2026-03-09T20:31:43.940 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-09T20:31:43.944 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-09T20:31:43.961 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/646 branch: squid description: orch/cephadm/workunits/{0-distro/centos_9.stream agent/on mon_election/connectivity task/test_rgw_multisite} email: null first_in_suite: false flavor: default job_id: '646' last_in_suite: false machine_type: vps name: kyr-2026-03-09_11:23:05-orch-squid-none-default-vps no_nested_subset: false os_type: centos os_version: 9.stream overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: global: mon election default strategy: 3 mgr: debug mgr: 20 debug ms: 1 mgr/cephadm/use_agent: true mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - MON_DOWN - mons down - mon down - out of quorum - CEPHADM_STRAY_DAEMON log-only-match: - CEPHADM_ sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 - scontext=system_u:system_r:getty_t:s0 workunit: branch: tt-squid sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - mon.a - mgr.a - osd.0 - - host.b - mon.b - mgr.b - osd.1 - - host.c - mon.c - osd.2 seed: 3443 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 targets: vm03.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBcyWM7Bnnsco/bAmiSl3C7FXXBu78H0lHd9jLu7o1HyEqveKIRsRDMCVGlm22ihhGtyFtAZlOuXoNc0dYl1Kf8= vm04.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPZ7vwElxYCkqoyOC8vHiMGeyR1xYNtGGGt5K94UCpv7tN5WYZVbwHnbiMXAtXaiGnVJfWlFoJYHQ9v/IypigfI= vm08.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNOJpmrJGIuo8TsWJqFICEbxmkHGzj7xQeJ15xa60MfGR2zVDVOHJdVQP/Mci7oww54SurcGut1pb1AX2vz74Z0= tasks: - pexec: all: - sudo dnf remove nvme-cli -y - sudo dnf install nvmetcli nvme-cli -y - install: null - cephadm: null - cephadm.shell: host.a: - ceph mgr module enable rgw - rgw_module.apply: specs: - rgw_realm: myrealm1 rgw_zone: myzone1 rgw_zonegroup: myzonegroup1 spec: rgw_frontend_port: 5500 - cephadm.shell: host.a: - 'set -e set -x while true; do TOKEN=$(ceph rgw realm tokens | jq -r ''.[0].token''); echo $TOKEN; if [ "$TOKEN" != "master zone has no endpoint" ]; then break; fi; sleep 5; done TOKENS=$(ceph rgw realm tokens) echo $TOKENS | jq --exit-status ''.[0].realm == "myrealm1"'' echo $TOKENS | jq --exit-status ''.[0].token'' TOKEN_JSON=$(ceph rgw realm tokens | jq -r ''.[0].token'' | base64 --decode) echo $TOKEN_JSON | jq --exit-status ''.realm_name == "myrealm1"'' echo $TOKEN_JSON | jq --exit-status ''.endpoint | test("http://.+:\\d+")'' echo $TOKEN_JSON | jq --exit-status ''.realm_id | test("^[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}$")'' echo $TOKEN_JSON | jq --exit-status ''.access_key'' echo $TOKEN_JSON | jq --exit-status ''.secret'' ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-09_11:23:05 tube: vps user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.611473 2026-03-09T20:31:43.962 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa; will attempt to use it 2026-03-09T20:31:43.962 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa/tasks 2026-03-09T20:31:43.962 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-09T20:31:43.962 INFO:teuthology.task.internal:Checking packages... 2026-03-09T20:31:43.963 INFO:teuthology.task.internal:Checking packages for os_type 'centos', flavor 'default' and ceph hash 'e911bdebe5c8faa3800735d1568fcdca65db60df' 2026-03-09T20:31:43.963 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-09T20:31:43.963 INFO:teuthology.packaging:ref: None 2026-03-09T20:31:43.963 INFO:teuthology.packaging:tag: None 2026-03-09T20:31:43.963 INFO:teuthology.packaging:branch: squid 2026-03-09T20:31:43.963 INFO:teuthology.packaging:sha1: e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:31:43.963 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&ref=squid 2026-03-09T20:31:44.774 INFO:teuthology.task.internal:Found packages for ceph version 19.2.3-678.ge911bdeb 2026-03-09T20:31:44.775 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-09T20:31:44.776 INFO:teuthology.task.internal:no buildpackages task found 2026-03-09T20:31:44.776 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-09T20:31:44.776 INFO:teuthology.task.internal:Saving configuration 2026-03-09T20:31:44.781 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-09T20:31:44.782 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-09T20:31:44.788 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm03.local', 'description': '/archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/646', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-09 20:30:04.131270', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:03', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBcyWM7Bnnsco/bAmiSl3C7FXXBu78H0lHd9jLu7o1HyEqveKIRsRDMCVGlm22ihhGtyFtAZlOuXoNc0dYl1Kf8='} 2026-03-09T20:31:44.793 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm04.local', 'description': '/archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/646', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-09 20:30:04.130999', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:04', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPZ7vwElxYCkqoyOC8vHiMGeyR1xYNtGGGt5K94UCpv7tN5WYZVbwHnbiMXAtXaiGnVJfWlFoJYHQ9v/IypigfI='} 2026-03-09T20:31:44.797 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm08.local', 'description': '/archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/646', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-09 20:30:04.130301', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:08', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNOJpmrJGIuo8TsWJqFICEbxmkHGzj7xQeJ15xa60MfGR2zVDVOHJdVQP/Mci7oww54SurcGut1pb1AX2vz74Z0='} 2026-03-09T20:31:44.797 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-09T20:31:44.798 INFO:teuthology.task.internal:roles: ubuntu@vm03.local - ['host.a', 'mon.a', 'mgr.a', 'osd.0'] 2026-03-09T20:31:44.798 INFO:teuthology.task.internal:roles: ubuntu@vm04.local - ['host.b', 'mon.b', 'mgr.b', 'osd.1'] 2026-03-09T20:31:44.798 INFO:teuthology.task.internal:roles: ubuntu@vm08.local - ['host.c', 'mon.c', 'osd.2'] 2026-03-09T20:31:44.798 INFO:teuthology.run_tasks:Running task console_log... 2026-03-09T20:31:44.803 DEBUG:teuthology.task.console_log:vm03 does not support IPMI; excluding 2026-03-09T20:31:44.807 DEBUG:teuthology.task.console_log:vm04 does not support IPMI; excluding 2026-03-09T20:31:44.812 DEBUG:teuthology.task.console_log:vm08 does not support IPMI; excluding 2026-03-09T20:31:44.812 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7f93b447a170>, signals=[15]) 2026-03-09T20:31:44.812 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-09T20:31:44.813 INFO:teuthology.task.internal:Opening connections... 2026-03-09T20:31:44.813 DEBUG:teuthology.task.internal:connecting to ubuntu@vm03.local 2026-03-09T20:31:44.813 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm03.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T20:31:44.874 DEBUG:teuthology.task.internal:connecting to ubuntu@vm04.local 2026-03-09T20:31:44.874 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm04.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T20:31:44.935 DEBUG:teuthology.task.internal:connecting to ubuntu@vm08.local 2026-03-09T20:31:44.936 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T20:31:44.997 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-09T20:31:44.999 DEBUG:teuthology.orchestra.run.vm03:> uname -m 2026-03-09T20:31:45.013 INFO:teuthology.orchestra.run.vm03.stdout:x86_64 2026-03-09T20:31:45.013 DEBUG:teuthology.orchestra.run.vm03:> cat /etc/os-release 2026-03-09T20:31:45.067 INFO:teuthology.orchestra.run.vm03.stdout:NAME="CentOS Stream" 2026-03-09T20:31:45.067 INFO:teuthology.orchestra.run.vm03.stdout:VERSION="9" 2026-03-09T20:31:45.067 INFO:teuthology.orchestra.run.vm03.stdout:ID="centos" 2026-03-09T20:31:45.067 INFO:teuthology.orchestra.run.vm03.stdout:ID_LIKE="rhel fedora" 2026-03-09T20:31:45.067 INFO:teuthology.orchestra.run.vm03.stdout:VERSION_ID="9" 2026-03-09T20:31:45.067 INFO:teuthology.orchestra.run.vm03.stdout:PLATFORM_ID="platform:el9" 2026-03-09T20:31:45.067 INFO:teuthology.orchestra.run.vm03.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T20:31:45.067 INFO:teuthology.orchestra.run.vm03.stdout:ANSI_COLOR="0;31" 2026-03-09T20:31:45.067 INFO:teuthology.orchestra.run.vm03.stdout:LOGO="fedora-logo-icon" 2026-03-09T20:31:45.067 INFO:teuthology.orchestra.run.vm03.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T20:31:45.067 INFO:teuthology.orchestra.run.vm03.stdout:HOME_URL="https://centos.org/" 2026-03-09T20:31:45.067 INFO:teuthology.orchestra.run.vm03.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T20:31:45.067 INFO:teuthology.orchestra.run.vm03.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T20:31:45.067 INFO:teuthology.orchestra.run.vm03.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T20:31:45.067 INFO:teuthology.lock.ops:Updating vm03.local on lock server 2026-03-09T20:31:45.072 DEBUG:teuthology.orchestra.run.vm04:> uname -m 2026-03-09T20:31:45.086 INFO:teuthology.orchestra.run.vm04.stdout:x86_64 2026-03-09T20:31:45.086 DEBUG:teuthology.orchestra.run.vm04:> cat /etc/os-release 2026-03-09T20:31:45.140 INFO:teuthology.orchestra.run.vm04.stdout:NAME="CentOS Stream" 2026-03-09T20:31:45.140 INFO:teuthology.orchestra.run.vm04.stdout:VERSION="9" 2026-03-09T20:31:45.140 INFO:teuthology.orchestra.run.vm04.stdout:ID="centos" 2026-03-09T20:31:45.140 INFO:teuthology.orchestra.run.vm04.stdout:ID_LIKE="rhel fedora" 2026-03-09T20:31:45.140 INFO:teuthology.orchestra.run.vm04.stdout:VERSION_ID="9" 2026-03-09T20:31:45.140 INFO:teuthology.orchestra.run.vm04.stdout:PLATFORM_ID="platform:el9" 2026-03-09T20:31:45.140 INFO:teuthology.orchestra.run.vm04.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T20:31:45.140 INFO:teuthology.orchestra.run.vm04.stdout:ANSI_COLOR="0;31" 2026-03-09T20:31:45.141 INFO:teuthology.orchestra.run.vm04.stdout:LOGO="fedora-logo-icon" 2026-03-09T20:31:45.141 INFO:teuthology.orchestra.run.vm04.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T20:31:45.141 INFO:teuthology.orchestra.run.vm04.stdout:HOME_URL="https://centos.org/" 2026-03-09T20:31:45.141 INFO:teuthology.orchestra.run.vm04.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T20:31:45.141 INFO:teuthology.orchestra.run.vm04.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T20:31:45.141 INFO:teuthology.orchestra.run.vm04.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T20:31:45.141 INFO:teuthology.lock.ops:Updating vm04.local on lock server 2026-03-09T20:31:45.145 DEBUG:teuthology.orchestra.run.vm08:> uname -m 2026-03-09T20:31:45.162 INFO:teuthology.orchestra.run.vm08.stdout:x86_64 2026-03-09T20:31:45.162 DEBUG:teuthology.orchestra.run.vm08:> cat /etc/os-release 2026-03-09T20:31:45.218 INFO:teuthology.orchestra.run.vm08.stdout:NAME="CentOS Stream" 2026-03-09T20:31:45.218 INFO:teuthology.orchestra.run.vm08.stdout:VERSION="9" 2026-03-09T20:31:45.218 INFO:teuthology.orchestra.run.vm08.stdout:ID="centos" 2026-03-09T20:31:45.218 INFO:teuthology.orchestra.run.vm08.stdout:ID_LIKE="rhel fedora" 2026-03-09T20:31:45.218 INFO:teuthology.orchestra.run.vm08.stdout:VERSION_ID="9" 2026-03-09T20:31:45.218 INFO:teuthology.orchestra.run.vm08.stdout:PLATFORM_ID="platform:el9" 2026-03-09T20:31:45.218 INFO:teuthology.orchestra.run.vm08.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T20:31:45.218 INFO:teuthology.orchestra.run.vm08.stdout:ANSI_COLOR="0;31" 2026-03-09T20:31:45.218 INFO:teuthology.orchestra.run.vm08.stdout:LOGO="fedora-logo-icon" 2026-03-09T20:31:45.218 INFO:teuthology.orchestra.run.vm08.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T20:31:45.218 INFO:teuthology.orchestra.run.vm08.stdout:HOME_URL="https://centos.org/" 2026-03-09T20:31:45.218 INFO:teuthology.orchestra.run.vm08.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T20:31:45.218 INFO:teuthology.orchestra.run.vm08.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T20:31:45.218 INFO:teuthology.orchestra.run.vm08.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T20:31:45.218 INFO:teuthology.lock.ops:Updating vm08.local on lock server 2026-03-09T20:31:45.223 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-09T20:31:45.224 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-09T20:31:45.225 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-09T20:31:45.225 DEBUG:teuthology.orchestra.run.vm03:> test '!' -e /home/ubuntu/cephtest 2026-03-09T20:31:45.227 DEBUG:teuthology.orchestra.run.vm04:> test '!' -e /home/ubuntu/cephtest 2026-03-09T20:31:45.229 DEBUG:teuthology.orchestra.run.vm08:> test '!' -e /home/ubuntu/cephtest 2026-03-09T20:31:45.274 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-09T20:31:45.286 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-09T20:31:45.286 DEBUG:teuthology.orchestra.run.vm03:> test -z $(ls -A /var/lib/ceph) 2026-03-09T20:31:45.287 DEBUG:teuthology.orchestra.run.vm04:> test -z $(ls -A /var/lib/ceph) 2026-03-09T20:31:45.289 DEBUG:teuthology.orchestra.run.vm08:> test -z $(ls -A /var/lib/ceph) 2026-03-09T20:31:45.302 INFO:teuthology.orchestra.run.vm04.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T20:31:45.304 INFO:teuthology.orchestra.run.vm03.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T20:31:45.334 INFO:teuthology.orchestra.run.vm08.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T20:31:45.334 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-09T20:31:45.342 DEBUG:teuthology.orchestra.run.vm03:> test -e /ceph-qa-ready 2026-03-09T20:31:45.362 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:31:45.557 DEBUG:teuthology.orchestra.run.vm04:> test -e /ceph-qa-ready 2026-03-09T20:31:45.572 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:31:45.770 DEBUG:teuthology.orchestra.run.vm08:> test -e /ceph-qa-ready 2026-03-09T20:31:45.787 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:31:45.972 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-09T20:31:45.973 INFO:teuthology.task.internal:Creating test directory... 2026-03-09T20:31:45.973 DEBUG:teuthology.orchestra.run.vm03:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T20:31:45.976 DEBUG:teuthology.orchestra.run.vm04:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T20:31:45.978 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T20:31:45.994 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-09T20:31:45.995 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-09T20:31:45.996 INFO:teuthology.task.internal:Creating archive directory... 2026-03-09T20:31:45.997 DEBUG:teuthology.orchestra.run.vm03:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T20:31:46.030 DEBUG:teuthology.orchestra.run.vm04:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T20:31:46.037 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T20:31:46.055 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-09T20:31:46.056 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-09T20:31:46.056 DEBUG:teuthology.orchestra.run.vm03:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T20:31:46.102 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:31:46.103 DEBUG:teuthology.orchestra.run.vm04:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T20:31:46.120 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:31:46.120 DEBUG:teuthology.orchestra.run.vm08:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T20:31:46.134 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:31:46.134 DEBUG:teuthology.orchestra.run.vm03:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T20:31:46.144 DEBUG:teuthology.orchestra.run.vm04:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T20:31:46.163 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T20:31:46.170 INFO:teuthology.orchestra.run.vm03.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T20:31:46.180 INFO:teuthology.orchestra.run.vm03.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T20:31:46.188 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T20:31:46.197 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T20:31:46.198 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T20:31:46.206 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T20:31:46.207 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-09T20:31:46.208 INFO:teuthology.task.internal:Configuring sudo... 2026-03-09T20:31:46.209 DEBUG:teuthology.orchestra.run.vm03:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T20:31:46.223 DEBUG:teuthology.orchestra.run.vm04:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T20:31:46.242 DEBUG:teuthology.orchestra.run.vm08:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T20:31:46.270 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-09T20:31:46.272 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-09T20:31:46.272 DEBUG:teuthology.orchestra.run.vm03:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T20:31:46.289 DEBUG:teuthology.orchestra.run.vm04:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T20:31:46.306 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T20:31:46.324 DEBUG:teuthology.orchestra.run.vm03:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T20:31:46.365 DEBUG:teuthology.orchestra.run.vm03:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T20:31:46.421 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T20:31:46.421 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T20:31:46.484 DEBUG:teuthology.orchestra.run.vm04:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T20:31:46.505 DEBUG:teuthology.orchestra.run.vm04:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T20:31:46.561 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T20:31:46.561 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T20:31:46.619 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T20:31:46.644 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T20:31:46.702 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:31:46.702 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T20:31:46.764 DEBUG:teuthology.orchestra.run.vm03:> sudo service rsyslog restart 2026-03-09T20:31:46.765 DEBUG:teuthology.orchestra.run.vm04:> sudo service rsyslog restart 2026-03-09T20:31:46.767 DEBUG:teuthology.orchestra.run.vm08:> sudo service rsyslog restart 2026-03-09T20:31:46.792 INFO:teuthology.orchestra.run.vm04.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T20:31:46.794 INFO:teuthology.orchestra.run.vm03.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T20:31:46.833 INFO:teuthology.orchestra.run.vm08.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T20:31:47.170 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-09T20:31:47.172 INFO:teuthology.task.internal:Starting timer... 2026-03-09T20:31:47.172 INFO:teuthology.run_tasks:Running task pcp... 2026-03-09T20:31:47.175 INFO:teuthology.run_tasks:Running task selinux... 2026-03-09T20:31:47.178 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0', 'scontext=system_u:system_r:getty_t:s0']} 2026-03-09T20:31:47.178 INFO:teuthology.task.selinux:Excluding vm03: VMs are not yet supported 2026-03-09T20:31:47.178 INFO:teuthology.task.selinux:Excluding vm04: VMs are not yet supported 2026-03-09T20:31:47.178 INFO:teuthology.task.selinux:Excluding vm08: VMs are not yet supported 2026-03-09T20:31:47.179 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-09T20:31:47.179 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-09T20:31:47.179 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-09T20:31:47.179 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-09T20:31:47.180 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-09T20:31:47.180 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-09T20:31:47.182 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-09T20:31:47.773 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-09T20:31:47.779 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-09T20:31:47.779 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventorytf4g_puw --limit vm03.local,vm04.local,vm08.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-09T20:33:50.327 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm03.local'), Remote(name='ubuntu@vm04.local'), Remote(name='ubuntu@vm08.local')] 2026-03-09T20:33:50.327 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm03.local' 2026-03-09T20:33:50.328 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm03.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T20:33:50.399 DEBUG:teuthology.orchestra.run.vm03:> true 2026-03-09T20:33:50.485 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm03.local' 2026-03-09T20:33:50.485 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm04.local' 2026-03-09T20:33:50.486 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm04.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T20:33:50.555 DEBUG:teuthology.orchestra.run.vm04:> true 2026-03-09T20:33:50.638 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm04.local' 2026-03-09T20:33:50.638 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm08.local' 2026-03-09T20:33:50.638 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T20:33:50.706 DEBUG:teuthology.orchestra.run.vm08:> true 2026-03-09T20:33:50.788 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm08.local' 2026-03-09T20:33:50.788 INFO:teuthology.run_tasks:Running task clock... 2026-03-09T20:33:50.791 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-09T20:33:50.791 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T20:33:50.791 DEBUG:teuthology.orchestra.run.vm03:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T20:33:50.794 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T20:33:50.794 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T20:33:50.798 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T20:33:50.798 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T20:33:50.834 INFO:teuthology.orchestra.run.vm04.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T20:33:50.835 INFO:teuthology.orchestra.run.vm03.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T20:33:50.849 INFO:teuthology.orchestra.run.vm04.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T20:33:50.851 INFO:teuthology.orchestra.run.vm03.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T20:33:50.880 INFO:teuthology.orchestra.run.vm04.stderr:sudo: ntpd: command not found 2026-03-09T20:33:50.881 INFO:teuthology.orchestra.run.vm08.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T20:33:50.888 INFO:teuthology.orchestra.run.vm03.stderr:sudo: ntpd: command not found 2026-03-09T20:33:50.893 INFO:teuthology.orchestra.run.vm04.stdout:506 Cannot talk to daemon 2026-03-09T20:33:50.898 INFO:teuthology.orchestra.run.vm08.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T20:33:50.904 INFO:teuthology.orchestra.run.vm03.stdout:506 Cannot talk to daemon 2026-03-09T20:33:50.909 INFO:teuthology.orchestra.run.vm04.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T20:33:50.925 INFO:teuthology.orchestra.run.vm03.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T20:33:50.925 INFO:teuthology.orchestra.run.vm08.stderr:sudo: ntpd: command not found 2026-03-09T20:33:50.926 INFO:teuthology.orchestra.run.vm04.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T20:33:50.941 INFO:teuthology.orchestra.run.vm08.stdout:506 Cannot talk to daemon 2026-03-09T20:33:50.947 INFO:teuthology.orchestra.run.vm03.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T20:33:50.958 INFO:teuthology.orchestra.run.vm08.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T20:33:50.976 INFO:teuthology.orchestra.run.vm08.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T20:33:50.976 INFO:teuthology.orchestra.run.vm04.stderr:bash: line 1: ntpq: command not found 2026-03-09T20:33:51.010 INFO:teuthology.orchestra.run.vm03.stderr:bash: line 1: ntpq: command not found 2026-03-09T20:33:51.031 INFO:teuthology.orchestra.run.vm08.stderr:bash: line 1: ntpq: command not found 2026-03-09T20:33:51.249 INFO:teuthology.orchestra.run.vm04.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T20:33:51.249 INFO:teuthology.orchestra.run.vm04.stdout:=============================================================================== 2026-03-09T20:33:51.249 INFO:teuthology.orchestra.run.vm04.stdout:^? router02.i-tk.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:33:51.249 INFO:teuthology.orchestra.run.vm04.stdout:^? ntp0.vikings.net 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:33:51.249 INFO:teuthology.orchestra.run.vm04.stdout:^? static.179.181.75.5.clie> 3 6 1 0 -1173us[-1173us] +/- 31ms 2026-03-09T20:33:51.249 INFO:teuthology.orchestra.run.vm04.stdout:^? cp.hypermediaa.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:33:51.250 INFO:teuthology.orchestra.run.vm03.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T20:33:51.250 INFO:teuthology.orchestra.run.vm03.stdout:=============================================================================== 2026-03-09T20:33:51.250 INFO:teuthology.orchestra.run.vm03.stdout:^? router02.i-tk.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:33:51.250 INFO:teuthology.orchestra.run.vm03.stdout:^? ntp0.vikings.net 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:33:51.250 INFO:teuthology.orchestra.run.vm03.stdout:^? static.179.181.75.5.clie> 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:33:51.250 INFO:teuthology.orchestra.run.vm03.stdout:^? cp.hypermediaa.de 2 6 1 0 -1801us[-1801us] +/- 18ms 2026-03-09T20:33:51.250 INFO:teuthology.orchestra.run.vm08.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T20:33:51.250 INFO:teuthology.orchestra.run.vm08.stdout:=============================================================================== 2026-03-09T20:33:51.250 INFO:teuthology.orchestra.run.vm08.stdout:^? cp.hypermediaa.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:33:51.250 INFO:teuthology.orchestra.run.vm08.stdout:^? router02.i-tk.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:33:51.250 INFO:teuthology.orchestra.run.vm08.stdout:^? ntp0.vikings.net 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:33:51.250 INFO:teuthology.orchestra.run.vm08.stdout:^? static.179.181.75.5.clie> 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-09T20:33:51.250 INFO:teuthology.run_tasks:Running task pexec... 2026-03-09T20:33:51.253 INFO:teuthology.task.pexec:Executing custom commands... 2026-03-09T20:33:51.253 DEBUG:teuthology.orchestra.run.vm03:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T20:33:51.253 DEBUG:teuthology.orchestra.run.vm04:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T20:33:51.253 DEBUG:teuthology.orchestra.run.vm08:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T20:33:51.255 DEBUG:teuthology.task.pexec:ubuntu@vm08.local< sudo dnf remove nvme-cli -y 2026-03-09T20:33:51.255 DEBUG:teuthology.task.pexec:ubuntu@vm08.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-09T20:33:51.256 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm08.local 2026-03-09T20:33:51.256 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T20:33:51.256 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-09T20:33:51.256 DEBUG:teuthology.task.pexec:ubuntu@vm03.local< sudo dnf remove nvme-cli -y 2026-03-09T20:33:51.256 DEBUG:teuthology.task.pexec:ubuntu@vm03.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-09T20:33:51.256 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm03.local 2026-03-09T20:33:51.256 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T20:33:51.256 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-09T20:33:51.292 DEBUG:teuthology.task.pexec:ubuntu@vm04.local< sudo dnf remove nvme-cli -y 2026-03-09T20:33:51.292 DEBUG:teuthology.task.pexec:ubuntu@vm04.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-09T20:33:51.292 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm04.local 2026-03-09T20:33:51.292 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T20:33:51.292 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-09T20:33:51.457 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: nvme-cli 2026-03-09T20:33:51.457 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:33:51.460 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:33:51.460 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:33:51.460 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:33:51.519 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: nvme-cli 2026-03-09T20:33:51.519 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:33:51.522 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:33:51.523 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:33:51.523 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:33:51.526 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: nvme-cli 2026-03-09T20:33:51.526 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:33:51.530 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:33:51.531 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:33:51.531 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:33:51.946 INFO:teuthology.orchestra.run.vm08.stdout:Last metadata expiration check: 0:01:18 ago on Mon 09 Mar 2026 08:32:33 PM UTC. 2026-03-09T20:33:52.049 INFO:teuthology.orchestra.run.vm04.stdout:Last metadata expiration check: 0:01:14 ago on Mon 09 Mar 2026 08:32:38 PM UTC. 2026-03-09T20:33:52.050 INFO:teuthology.orchestra.run.vm03.stdout:Last metadata expiration check: 0:01:13 ago on Mon 09 Mar 2026 08:32:39 PM UTC. 2026-03-09T20:33:52.073 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout: Package Architecture Version Repository Size 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout:Installing: 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout:Installing dependencies: 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout:Install 6 Packages 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout:Total download size: 2.3 M 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout:Installed size: 11 M 2026-03-09T20:33:52.074 INFO:teuthology.orchestra.run.vm08.stdout:Downloading Packages: 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout: Package Architecture Version Repository Size 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout:Installing: 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout:Installing dependencies: 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout:Install 6 Packages 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:33:52.183 INFO:teuthology.orchestra.run.vm04.stdout:Total download size: 2.3 M 2026-03-09T20:33:52.184 INFO:teuthology.orchestra.run.vm04.stdout:Installed size: 11 M 2026-03-09T20:33:52.184 INFO:teuthology.orchestra.run.vm04.stdout:Downloading Packages: 2026-03-09T20:33:52.184 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout: Package Architecture Version Repository Size 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout:Installing: 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout:Installing dependencies: 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout:Install 6 Packages 2026-03-09T20:33:52.185 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:33:52.186 INFO:teuthology.orchestra.run.vm03.stdout:Total download size: 2.3 M 2026-03-09T20:33:52.186 INFO:teuthology.orchestra.run.vm03.stdout:Installed size: 11 M 2026-03-09T20:33:52.186 INFO:teuthology.orchestra.run.vm03.stdout:Downloading Packages: 2026-03-09T20:33:52.325 INFO:teuthology.orchestra.run.vm08.stdout:(1/6): python3-configshell-1.1.30-1.el9.noarch. 631 kB/s | 72 kB 00:00 2026-03-09T20:33:52.406 INFO:teuthology.orchestra.run.vm08.stdout:(2/6): python3-kmod-0.9-32.el9.x86_64.rpm 1.0 MB/s | 84 kB 00:00 2026-03-09T20:33:52.474 INFO:teuthology.orchestra.run.vm08.stdout:(3/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 2.2 MB/s | 150 kB 00:00 2026-03-09T20:33:52.506 INFO:teuthology.orchestra.run.vm08.stdout:(4/6): nvmetcli-0.8-3.el9.noarch.rpm 149 kB/s | 44 kB 00:00 2026-03-09T20:33:52.598 INFO:teuthology.orchestra.run.vm08.stdout:(5/6): nvme-cli-2.16-1.el9.x86_64.rpm 3.0 MB/s | 1.2 MB 00:00 2026-03-09T20:33:52.599 INFO:teuthology.orchestra.run.vm04.stdout:(1/6): nvmetcli-0.8-3.el9.noarch.rpm 156 kB/s | 44 kB 00:00 2026-03-09T20:33:52.600 INFO:teuthology.orchestra.run.vm04.stdout:(2/6): python3-configshell-1.1.30-1.el9.noarch. 255 kB/s | 72 kB 00:00 2026-03-09T20:33:52.641 INFO:teuthology.orchestra.run.vm08.stdout:(6/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 4.9 MB/s | 837 kB 00:00 2026-03-09T20:33:52.641 INFO:teuthology.orchestra.run.vm08.stdout:-------------------------------------------------------------------------------- 2026-03-09T20:33:52.641 INFO:teuthology.orchestra.run.vm08.stdout:Total 4.1 MB/s | 2.3 MB 00:00 2026-03-09T20:33:52.723 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:33:52.726 INFO:teuthology.orchestra.run.vm04.stdout:(3/6): nvme-cli-2.16-1.el9.x86_64.rpm 2.8 MB/s | 1.2 MB 00:00 2026-03-09T20:33:52.726 INFO:teuthology.orchestra.run.vm04.stdout:(4/6): python3-kmod-0.9-32.el9.x86_64.rpm 660 kB/s | 84 kB 00:00 2026-03-09T20:33:52.729 INFO:teuthology.orchestra.run.vm04.stdout:(5/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.1 MB/s | 150 kB 00:00 2026-03-09T20:33:52.735 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:33:52.735 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:33:52.796 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:33:52.796 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:33:52.867 INFO:teuthology.orchestra.run.vm04.stdout:(6/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 5.9 MB/s | 837 kB 00:00 2026-03-09T20:33:52.871 INFO:teuthology.orchestra.run.vm04.stdout:-------------------------------------------------------------------------------- 2026-03-09T20:33:52.871 INFO:teuthology.orchestra.run.vm04.stdout:Total 3.4 MB/s | 2.3 MB 00:00 2026-03-09T20:33:52.928 INFO:teuthology.orchestra.run.vm03.stdout:(1/6): python3-configshell-1.1.30-1.el9.noarch. 343 kB/s | 72 kB 00:00 2026-03-09T20:33:52.929 INFO:teuthology.orchestra.run.vm03.stdout:(2/6): nvmetcli-0.8-3.el9.noarch.rpm 208 kB/s | 44 kB 00:00 2026-03-09T20:33:52.957 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T20:33:52.968 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T20:33:52.968 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T20:33:52.992 INFO:teuthology.orchestra.run.vm03.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 1.3 MB/s | 84 kB 00:00 2026-03-09T20:33:53.039 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T20:33:53.040 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T20:33:53.045 INFO:teuthology.orchestra.run.vm03.stdout:(4/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.3 MB/s | 150 kB 00:00 2026-03-09T20:33:53.066 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:33:53.081 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-09T20:33:53.093 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-09T20:33:53.103 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T20:33:53.164 INFO:teuthology.orchestra.run.vm03.stdout:(5/6): nvme-cli-2.16-1.el9.x86_64.rpm 2.6 MB/s | 1.2 MB 00:00 2026-03-09T20:33:53.172 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T20:33:53.174 INFO:teuthology.orchestra.run.vm08.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T20:33:53.238 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T20:33:53.318 INFO:teuthology.orchestra.run.vm03.stdout:(6/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 2.5 MB/s | 837 kB 00:00 2026-03-09T20:33:53.318 INFO:teuthology.orchestra.run.vm03.stdout:-------------------------------------------------------------------------------- 2026-03-09T20:33:53.318 INFO:teuthology.orchestra.run.vm03.stdout:Total 2.0 MB/s | 2.3 MB 00:01 2026-03-09T20:33:53.327 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-09T20:33:53.345 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-09T20:33:53.358 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T20:33:53.368 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T20:33:53.376 INFO:teuthology.orchestra.run.vm04.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T20:33:53.390 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T20:33:53.395 INFO:teuthology.orchestra.run.vm08.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T20:33:53.400 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T20:33:53.411 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T20:33:53.411 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T20:33:53.471 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T20:33:53.471 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T20:33:53.604 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T20:33:53.610 INFO:teuthology.orchestra.run.vm04.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T20:33:53.667 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T20:33:53.679 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-09T20:33:53.696 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-09T20:33:53.704 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T20:33:53.715 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T20:33:53.718 INFO:teuthology.orchestra.run.vm03.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T20:33:53.869 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T20:33:53.869 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T20:33:53.869 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:33:53.933 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T20:33:53.941 INFO:teuthology.orchestra.run.vm03.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T20:33:54.043 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T20:33:54.043 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T20:33:54.043 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:33:54.348 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T20:33:54.348 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T20:33:54.348 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:33:54.555 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-09T20:33:54.556 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-09T20:33:54.556 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T20:33:54.556 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T20:33:54.556 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-09T20:33:54.660 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-09T20:33:54.660 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:33:54.660 INFO:teuthology.orchestra.run.vm08.stdout:Installed: 2026-03-09T20:33:54.660 INFO:teuthology.orchestra.run.vm08.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T20:33:54.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T20:33:54.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T20:33:54.660 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:33:54.660 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:33:54.732 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-09T20:33:54.733 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-09T20:33:54.733 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T20:33:54.733 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T20:33:54.733 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-09T20:33:54.778 DEBUG:teuthology.parallel:result is None 2026-03-09T20:33:54.844 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-09T20:33:54.844 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:33:54.844 INFO:teuthology.orchestra.run.vm04.stdout:Installed: 2026-03-09T20:33:54.844 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T20:33:54.844 INFO:teuthology.orchestra.run.vm04.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T20:33:54.844 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T20:33:54.844 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:33:54.844 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:33:54.969 DEBUG:teuthology.parallel:result is None 2026-03-09T20:33:54.995 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-09T20:33:54.995 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-09T20:33:54.995 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T20:33:54.995 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T20:33:54.995 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-09T20:33:55.092 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-09T20:33:55.092 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:33:55.092 INFO:teuthology.orchestra.run.vm03.stdout:Installed: 2026-03-09T20:33:55.092 INFO:teuthology.orchestra.run.vm03.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T20:33:55.092 INFO:teuthology.orchestra.run.vm03.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T20:33:55.092 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T20:33:55.092 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:33:55.092 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:33:55.179 DEBUG:teuthology.parallel:result is None 2026-03-09T20:33:55.179 INFO:teuthology.run_tasks:Running task install... 2026-03-09T20:33:55.181 DEBUG:teuthology.task.install:project ceph 2026-03-09T20:33:55.181 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'}, 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-09T20:33:55.181 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-09T20:33:55.181 INFO:teuthology.task.install:Using flavor: default 2026-03-09T20:33:55.184 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-09T20:33:55.184 INFO:teuthology.task.install:extra packages: [] 2026-03-09T20:33:55.184 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T20:33:55.184 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:33:55.184 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T20:33:55.185 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:33:55.185 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T20:33:55.185 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:33:55.852 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T20:33:55.852 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T20:33:55.886 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T20:33:55.886 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T20:33:55.886 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T20:33:55.886 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T20:33:56.368 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T20:33:56.368 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T20:33:56.368 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T20:33:56.394 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T20:33:56.395 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T20:33:56.395 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T20:33:56.406 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T20:33:56.406 DEBUG:teuthology.orchestra.run.vm03:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T20:33:56.409 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T20:33:56.410 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:33:56.410 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T20:33:56.435 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T20:33:56.435 DEBUG:teuthology.orchestra.run.vm04:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T20:33:56.451 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T20:33:56.451 DEBUG:teuthology.orchestra.run.vm08:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T20:33:56.486 DEBUG:teuthology.orchestra.run.vm03:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T20:33:56.506 DEBUG:teuthology.orchestra.run.vm04:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T20:33:56.529 DEBUG:teuthology.orchestra.run.vm08:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T20:33:56.585 DEBUG:teuthology.orchestra.run.vm03:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T20:33:56.599 DEBUG:teuthology.orchestra.run.vm04:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T20:33:56.619 DEBUG:teuthology.orchestra.run.vm08:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T20:33:56.626 INFO:teuthology.orchestra.run.vm03.stdout:check_obsoletes = 1 2026-03-09T20:33:56.627 DEBUG:teuthology.orchestra.run.vm03:> sudo yum clean all 2026-03-09T20:33:56.633 INFO:teuthology.orchestra.run.vm04.stdout:check_obsoletes = 1 2026-03-09T20:33:56.635 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean all 2026-03-09T20:33:56.655 INFO:teuthology.orchestra.run.vm08.stdout:check_obsoletes = 1 2026-03-09T20:33:56.657 DEBUG:teuthology.orchestra.run.vm08:> sudo yum clean all 2026-03-09T20:33:56.859 INFO:teuthology.orchestra.run.vm03.stdout:41 files removed 2026-03-09T20:33:56.867 INFO:teuthology.orchestra.run.vm08.stdout:41 files removed 2026-03-09T20:33:56.871 INFO:teuthology.orchestra.run.vm04.stdout:41 files removed 2026-03-09T20:33:56.891 DEBUG:teuthology.orchestra.run.vm08:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T20:33:56.896 DEBUG:teuthology.orchestra.run.vm04:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T20:33:56.897 DEBUG:teuthology.orchestra.run.vm03:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T20:33:58.250 INFO:teuthology.orchestra.run.vm08.stdout:ceph packages for x86_64 73 kB/s | 84 kB 00:01 2026-03-09T20:33:58.261 INFO:teuthology.orchestra.run.vm04.stdout:ceph packages for x86_64 73 kB/s | 84 kB 00:01 2026-03-09T20:33:58.280 INFO:teuthology.orchestra.run.vm03.stdout:ceph packages for x86_64 72 kB/s | 84 kB 00:01 2026-03-09T20:33:59.199 INFO:teuthology.orchestra.run.vm04.stdout:ceph noarch packages 13 kB/s | 12 kB 00:00 2026-03-09T20:33:59.227 INFO:teuthology.orchestra.run.vm08.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-09T20:33:59.281 INFO:teuthology.orchestra.run.vm03.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-09T20:34:00.152 INFO:teuthology.orchestra.run.vm04.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T20:34:00.189 INFO:teuthology.orchestra.run.vm08.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T20:34:00.259 INFO:teuthology.orchestra.run.vm03.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T20:34:00.748 INFO:teuthology.orchestra.run.vm08.stdout:CentOS Stream 9 - BaseOS 17 MB/s | 8.9 MB 00:00 2026-03-09T20:34:01.597 INFO:teuthology.orchestra.run.vm03.stdout:CentOS Stream 9 - BaseOS 6.8 MB/s | 8.9 MB 00:01 2026-03-09T20:34:01.631 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - BaseOS 6.1 MB/s | 8.9 MB 00:01 2026-03-09T20:34:02.876 INFO:teuthology.orchestra.run.vm08.stdout:CentOS Stream 9 - AppStream 23 MB/s | 27 MB 00:01 2026-03-09T20:34:03.917 INFO:teuthology.orchestra.run.vm03.stdout:CentOS Stream 9 - AppStream 18 MB/s | 27 MB 00:01 2026-03-09T20:34:04.834 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - AppStream 11 MB/s | 27 MB 00:02 2026-03-09T20:34:08.996 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - CRB 6.2 MB/s | 8.0 MB 00:01 2026-03-09T20:34:09.477 INFO:teuthology.orchestra.run.vm03.stdout:CentOS Stream 9 - CRB 3.8 MB/s | 8.0 MB 00:02 2026-03-09T20:34:10.297 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - Extras packages 46 kB/s | 20 kB 00:00 2026-03-09T20:34:10.771 INFO:teuthology.orchestra.run.vm04.stdout:Extra Packages for Enterprise Linux 52 MB/s | 20 MB 00:00 2026-03-09T20:34:11.096 INFO:teuthology.orchestra.run.vm03.stdout:CentOS Stream 9 - Extras packages 26 kB/s | 20 kB 00:00 2026-03-09T20:34:11.598 INFO:teuthology.orchestra.run.vm08.stdout:CentOS Stream 9 - CRB 1.5 MB/s | 8.0 MB 00:05 2026-03-09T20:34:11.979 INFO:teuthology.orchestra.run.vm03.stdout:Extra Packages for Enterprise Linux 26 MB/s | 20 MB 00:00 2026-03-09T20:34:13.918 INFO:teuthology.orchestra.run.vm08.stdout:CentOS Stream 9 - Extras packages 14 kB/s | 20 kB 00:01 2026-03-09T20:34:14.700 INFO:teuthology.orchestra.run.vm08.stdout:Extra Packages for Enterprise Linux 29 MB/s | 20 MB 00:00 2026-03-09T20:34:15.465 INFO:teuthology.orchestra.run.vm04.stdout:lab-extras 64 kB/s | 50 kB 00:00 2026-03-09T20:34:16.750 INFO:teuthology.orchestra.run.vm03.stdout:lab-extras 64 kB/s | 50 kB 00:00 2026-03-09T20:34:16.903 INFO:teuthology.orchestra.run.vm04.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T20:34:16.903 INFO:teuthology.orchestra.run.vm04.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T20:34:16.907 INFO:teuthology.orchestra.run.vm04.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T20:34:16.908 INFO:teuthology.orchestra.run.vm04.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T20:34:16.936 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout:====================================================================================== 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout:====================================================================================== 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout:Installing: 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T20:34:16.941 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout:Upgrading: 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout:Installing dependencies: 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T20:34:16.942 INFO:teuthology.orchestra.run.vm04.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T20:34:16.943 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout:Installing weak dependencies: 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout:====================================================================================== 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout:Install 134 Packages 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout:Upgrade 2 Packages 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout:Total download size: 210 M 2026-03-09T20:34:16.944 INFO:teuthology.orchestra.run.vm04.stdout:Downloading Packages: 2026-03-09T20:34:18.188 INFO:teuthology.orchestra.run.vm03.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T20:34:18.188 INFO:teuthology.orchestra.run.vm03.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T20:34:18.193 INFO:teuthology.orchestra.run.vm03.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T20:34:18.193 INFO:teuthology.orchestra.run.vm03.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T20:34:18.230 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout:====================================================================================== 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout:====================================================================================== 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout:Installing: 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout:Upgrading: 2026-03-09T20:34:18.235 INFO:teuthology.orchestra.run.vm03.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout:Installing dependencies: 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T20:34:18.236 INFO:teuthology.orchestra.run.vm03.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T20:34:18.237 INFO:teuthology.orchestra.run.vm03.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout:Installing weak dependencies: 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout:====================================================================================== 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout:Install 134 Packages 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout:Upgrade 2 Packages 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout:Total download size: 210 M 2026-03-09T20:34:18.238 INFO:teuthology.orchestra.run.vm03.stdout:Downloading Packages: 2026-03-09T20:34:18.641 INFO:teuthology.orchestra.run.vm04.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 14 kB/s | 6.5 kB 00:00 2026-03-09T20:34:19.381 INFO:teuthology.orchestra.run.vm08.stdout:lab-extras 64 kB/s | 50 kB 00:00 2026-03-09T20:34:19.688 INFO:teuthology.orchestra.run.vm04.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.1 MB/s | 1.2 MB 00:01 2026-03-09T20:34:19.807 INFO:teuthology.orchestra.run.vm04.stdout:(3/136): ceph-immutable-object-cache-19.2.3-678 1.2 MB/s | 145 kB 00:00 2026-03-09T20:34:20.067 INFO:teuthology.orchestra.run.vm03.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 13 kB/s | 6.5 kB 00:00 2026-03-09T20:34:20.747 INFO:teuthology.orchestra.run.vm04.stdout:(4/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 2.6 MB/s | 2.4 MB 00:00 2026-03-09T20:34:20.907 INFO:teuthology.orchestra.run.vm08.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T20:34:20.908 INFO:teuthology.orchestra.run.vm08.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T20:34:20.912 INFO:teuthology.orchestra.run.vm08.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T20:34:20.912 INFO:teuthology.orchestra.run.vm08.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T20:34:20.941 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout:====================================================================================== 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout:====================================================================================== 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout:Installing: 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T20:34:20.946 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout:Upgrading: 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout:Installing dependencies: 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T20:34:20.947 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T20:34:20.948 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T20:34:20.949 INFO:teuthology.orchestra.run.vm08.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T20:34:20.950 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T20:34:20.950 INFO:teuthology.orchestra.run.vm08.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T20:34:20.950 INFO:teuthology.orchestra.run.vm08.stdout:Installing weak dependencies: 2026-03-09T20:34:20.950 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T20:34:20.950 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:34:20.950 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:34:20.950 INFO:teuthology.orchestra.run.vm08.stdout:====================================================================================== 2026-03-09T20:34:20.950 INFO:teuthology.orchestra.run.vm08.stdout:Install 134 Packages 2026-03-09T20:34:20.950 INFO:teuthology.orchestra.run.vm08.stdout:Upgrade 2 Packages 2026-03-09T20:34:20.950 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:34:20.950 INFO:teuthology.orchestra.run.vm08.stdout:Total download size: 210 M 2026-03-09T20:34:20.950 INFO:teuthology.orchestra.run.vm08.stdout:Downloading Packages: 2026-03-09T20:34:20.990 INFO:teuthology.orchestra.run.vm04.stdout:(5/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 2.0 MB/s | 5.5 MB 00:02 2026-03-09T20:34:21.128 INFO:teuthology.orchestra.run.vm03.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.1 MB/s | 1.2 MB 00:01 2026-03-09T20:34:21.228 INFO:teuthology.orchestra.run.vm04.stdout:(6/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 2.2 MB/s | 1.1 MB 00:00 2026-03-09T20:34:21.249 INFO:teuthology.orchestra.run.vm03.stdout:(3/136): ceph-immutable-object-cache-19.2.3-678 1.2 MB/s | 145 kB 00:00 2026-03-09T20:34:22.396 INFO:teuthology.orchestra.run.vm04.stdout:(7/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 3.4 MB/s | 4.7 MB 00:01 2026-03-09T20:34:22.580 INFO:teuthology.orchestra.run.vm08.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 14 kB/s | 6.5 kB 00:00 2026-03-09T20:34:22.775 INFO:teuthology.orchestra.run.vm03.stdout:(4/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 1.6 MB/s | 2.4 MB 00:01 2026-03-09T20:34:22.801 INFO:teuthology.orchestra.run.vm03.stdout:(5/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 1.7 MB/s | 5.5 MB 00:03 2026-03-09T20:34:23.363 INFO:teuthology.orchestra.run.vm03.stdout:(6/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 1.8 MB/s | 1.1 MB 00:00 2026-03-09T20:34:23.488 INFO:teuthology.orchestra.run.vm08.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.3 MB/s | 1.2 MB 00:00 2026-03-09T20:34:23.604 INFO:teuthology.orchestra.run.vm08.stdout:(3/136): ceph-immutable-object-cache-19.2.3-678 1.2 MB/s | 145 kB 00:00 2026-03-09T20:34:24.510 INFO:teuthology.orchestra.run.vm08.stdout:(4/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 2.7 MB/s | 2.4 MB 00:00 2026-03-09T20:34:24.583 INFO:teuthology.orchestra.run.vm03.stdout:(7/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 2.7 MB/s | 4.7 MB 00:01 2026-03-09T20:34:24.775 INFO:teuthology.orchestra.run.vm04.stdout:(8/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9. 4.5 MB/s | 11 MB 00:02 2026-03-09T20:34:24.781 INFO:teuthology.orchestra.run.vm08.stdout:(5/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 2.1 MB/s | 5.5 MB 00:02 2026-03-09T20:34:24.852 INFO:teuthology.orchestra.run.vm08.stdout:(6/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 3.1 MB/s | 1.1 MB 00:00 2026-03-09T20:34:24.915 INFO:teuthology.orchestra.run.vm04.stdout:(9/136): ceph-selinux-19.2.3-678.ge911bdeb.el9. 179 kB/s | 25 kB 00:00 2026-03-09T20:34:26.400 INFO:teuthology.orchestra.run.vm04.stdout:(10/136): ceph-common-19.2.3-678.ge911bdeb.el9. 2.6 MB/s | 22 MB 00:08 2026-03-09T20:34:26.401 INFO:teuthology.orchestra.run.vm03.stdout:(8/136): ceph-common-19.2.3-678.ge911bdeb.el9.x 3.2 MB/s | 22 MB 00:06 2026-03-09T20:34:26.510 INFO:teuthology.orchestra.run.vm03.stdout:(9/136): ceph-selinux-19.2.3-678.ge911bdeb.el9. 229 kB/s | 25 kB 00:00 2026-03-09T20:34:26.522 INFO:teuthology.orchestra.run.vm04.stdout:(11/136): libcephfs-devel-19.2.3-678.ge911bdeb. 276 kB/s | 34 kB 00:00 2026-03-09T20:34:26.740 INFO:teuthology.orchestra.run.vm08.stdout:(7/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 2.4 MB/s | 4.7 MB 00:01 2026-03-09T20:34:26.853 INFO:teuthology.orchestra.run.vm04.stdout:(12/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 3.0 MB/s | 17 MB 00:05 2026-03-09T20:34:26.889 INFO:teuthology.orchestra.run.vm04.stdout:(13/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 2.7 MB/s | 1.0 MB 00:00 2026-03-09T20:34:26.976 INFO:teuthology.orchestra.run.vm04.stdout:(14/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.3 MB/s | 163 kB 00:00 2026-03-09T20:34:27.010 INFO:teuthology.orchestra.run.vm04.stdout:(15/136): librados-devel-19.2.3-678.ge911bdeb.e 1.0 MB/s | 127 kB 00:00 2026-03-09T20:34:27.052 INFO:teuthology.orchestra.run.vm08.stdout:(8/136): ceph-common-19.2.3-678.ge911bdeb.el9.x 4.4 MB/s | 22 MB 00:04 2026-03-09T20:34:27.166 INFO:teuthology.orchestra.run.vm08.stdout:(9/136): ceph-selinux-19.2.3-678.ge911bdeb.el9. 221 kB/s | 25 kB 00:00 2026-03-09T20:34:27.213 INFO:teuthology.orchestra.run.vm04.stdout:(16/136): libradosstriper1-19.2.3-678.ge911bdeb 2.1 MB/s | 503 kB 00:00 2026-03-09T20:34:27.333 INFO:teuthology.orchestra.run.vm04.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 377 kB/s | 45 kB 00:00 2026-03-09T20:34:27.457 INFO:teuthology.orchestra.run.vm04.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 1.1 MB/s | 142 kB 00:00 2026-03-09T20:34:27.578 INFO:teuthology.orchestra.run.vm04.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 1.3 MB/s | 165 kB 00:00 2026-03-09T20:34:27.699 INFO:teuthology.orchestra.run.vm04.stdout:(20/136): python3-rados-19.2.3-678.ge911bdeb.el 2.6 MB/s | 323 kB 00:00 2026-03-09T20:34:27.820 INFO:teuthology.orchestra.run.vm04.stdout:(21/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 303 kB 00:00 2026-03-09T20:34:27.939 INFO:teuthology.orchestra.run.vm04.stdout:(22/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 837 kB/s | 100 kB 00:00 2026-03-09T20:34:28.059 INFO:teuthology.orchestra.run.vm04.stdout:(23/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 711 kB/s | 85 kB 00:00 2026-03-09T20:34:28.390 INFO:teuthology.orchestra.run.vm03.stdout:(10/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9 2.8 MB/s | 11 MB 00:03 2026-03-09T20:34:28.511 INFO:teuthology.orchestra.run.vm03.stdout:(11/136): libcephfs-devel-19.2.3-678.ge911bdeb. 279 kB/s | 34 kB 00:00 2026-03-09T20:34:28.871 INFO:teuthology.orchestra.run.vm03.stdout:(12/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 2.7 MB/s | 1.0 MB 00:00 2026-03-09T20:34:28.905 INFO:teuthology.orchestra.run.vm04.stdout:(24/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 2.9 MB/s | 5.4 MB 00:01 2026-03-09T20:34:28.941 INFO:teuthology.orchestra.run.vm08.stdout:(10/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 4.2 MB/s | 17 MB 00:04 2026-03-09T20:34:28.992 INFO:teuthology.orchestra.run.vm03.stdout:(13/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.3 MB/s | 163 kB 00:00 2026-03-09T20:34:29.026 INFO:teuthology.orchestra.run.vm04.stdout:(25/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.4 MB/s | 171 kB 00:00 2026-03-09T20:34:29.052 INFO:teuthology.orchestra.run.vm08.stdout:(11/136): libcephfs-devel-19.2.3-678.ge911bdeb. 298 kB/s | 34 kB 00:00 2026-03-09T20:34:29.110 INFO:teuthology.orchestra.run.vm04.stdout:(26/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 3.0 MB/s | 3.1 MB 00:01 2026-03-09T20:34:29.113 INFO:teuthology.orchestra.run.vm03.stdout:(14/136): librados-devel-19.2.3-678.ge911bdeb.e 1.0 MB/s | 127 kB 00:00 2026-03-09T20:34:29.146 INFO:teuthology.orchestra.run.vm04.stdout:(27/136): ceph-grafana-dashboards-19.2.3-678.ge 261 kB/s | 31 kB 00:00 2026-03-09T20:34:29.231 INFO:teuthology.orchestra.run.vm04.stdout:(28/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.2 MB/s | 150 kB 00:00 2026-03-09T20:34:29.284 INFO:teuthology.orchestra.run.vm08.stdout:(12/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 4.2 MB/s | 1.0 MB 00:00 2026-03-09T20:34:29.354 INFO:teuthology.orchestra.run.vm03.stdout:(15/136): libradosstriper1-19.2.3-678.ge911bdeb 2.0 MB/s | 503 kB 00:00 2026-03-09T20:34:29.398 INFO:teuthology.orchestra.run.vm08.stdout:(13/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.4 MB/s | 163 kB 00:00 2026-03-09T20:34:29.512 INFO:teuthology.orchestra.run.vm08.stdout:(14/136): librados-devel-19.2.3-678.ge911bdeb.e 1.1 MB/s | 127 kB 00:00 2026-03-09T20:34:29.629 INFO:teuthology.orchestra.run.vm08.stdout:(15/136): libradosstriper1-19.2.3-678.ge911bdeb 4.2 MB/s | 503 kB 00:00 2026-03-09T20:34:30.311 INFO:teuthology.orchestra.run.vm03.stdout:(16/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 2.5 MB/s | 17 MB 00:06 2026-03-09T20:34:30.429 INFO:teuthology.orchestra.run.vm03.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 380 kB/s | 45 kB 00:00 2026-03-09T20:34:30.601 INFO:teuthology.orchestra.run.vm03.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 831 kB/s | 142 kB 00:00 2026-03-09T20:34:30.603 INFO:teuthology.orchestra.run.vm08.stdout:(16/136): ceph-test-19.2.3-678.ge911bdeb.el9.x8 14 MB/s | 50 MB 00:03 2026-03-09T20:34:30.690 INFO:teuthology.orchestra.run.vm04.stdout:(29/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 2.5 MB/s | 3.8 MB 00:01 2026-03-09T20:34:30.717 INFO:teuthology.orchestra.run.vm08.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 397 kB/s | 45 kB 00:00 2026-03-09T20:34:30.720 INFO:teuthology.orchestra.run.vm03.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 1.4 MB/s | 165 kB 00:00 2026-03-09T20:34:30.814 INFO:teuthology.orchestra.run.vm04.stdout:(30/136): ceph-mgr-modules-core-19.2.3-678.ge91 2.0 MB/s | 253 kB 00:00 2026-03-09T20:34:30.832 INFO:teuthology.orchestra.run.vm08.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 1.2 MB/s | 142 kB 00:00 2026-03-09T20:34:30.935 INFO:teuthology.orchestra.run.vm04.stdout:(31/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 408 kB/s | 49 kB 00:00 2026-03-09T20:34:30.947 INFO:teuthology.orchestra.run.vm08.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 1.4 MB/s | 165 kB 00:00 2026-03-09T20:34:30.955 INFO:teuthology.orchestra.run.vm03.stdout:(20/136): python3-rados-19.2.3-678.ge911bdeb.el 1.3 MB/s | 323 kB 00:00 2026-03-09T20:34:31.055 INFO:teuthology.orchestra.run.vm04.stdout:(32/136): ceph-prometheus-alerts-19.2.3-678.ge9 140 kB/s | 17 kB 00:00 2026-03-09T20:34:31.064 INFO:teuthology.orchestra.run.vm08.stdout:(20/136): python3-rados-19.2.3-678.ge911bdeb.el 2.7 MB/s | 323 kB 00:00 2026-03-09T20:34:31.141 INFO:teuthology.orchestra.run.vm08.stdout:(21/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9 2.4 MB/s | 11 MB 00:04 2026-03-09T20:34:31.157 INFO:teuthology.orchestra.run.vm08.stdout:(22/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 3.5 MB/s | 5.4 MB 00:01 2026-03-09T20:34:31.180 INFO:teuthology.orchestra.run.vm08.stdout:(23/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.5 MB/s | 303 kB 00:00 2026-03-09T20:34:31.189 INFO:teuthology.orchestra.run.vm03.stdout:(21/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 1.3 MB/s | 303 kB 00:00 2026-03-09T20:34:31.258 INFO:teuthology.orchestra.run.vm08.stdout:(24/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 859 kB/s | 100 kB 00:00 2026-03-09T20:34:31.270 INFO:teuthology.orchestra.run.vm08.stdout:(25/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 754 kB/s | 85 kB 00:00 2026-03-09T20:34:31.295 INFO:teuthology.orchestra.run.vm04.stdout:(33/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 1.2 MB/s | 299 kB 00:00 2026-03-09T20:34:31.314 INFO:teuthology.orchestra.run.vm03.stdout:(22/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 801 kB/s | 100 kB 00:00 2026-03-09T20:34:31.374 INFO:teuthology.orchestra.run.vm08.stdout:(26/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.4 MB/s | 171 kB 00:00 2026-03-09T20:34:31.383 INFO:teuthology.orchestra.run.vm08.stdout:(27/136): ceph-grafana-dashboards-19.2.3-678.ge 278 kB/s | 31 kB 00:00 2026-03-09T20:34:31.434 INFO:teuthology.orchestra.run.vm03.stdout:(23/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 708 kB/s | 85 kB 00:00 2026-03-09T20:34:31.496 INFO:teuthology.orchestra.run.vm08.stdout:(28/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.2 MB/s | 150 kB 00:00 2026-03-09T20:34:31.535 INFO:teuthology.orchestra.run.vm08.stdout:(29/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 8.8 MB/s | 3.1 MB 00:00 2026-03-09T20:34:31.569 INFO:teuthology.orchestra.run.vm04.stdout:(34/136): ceph-mgr-diskprediction-local-19.2.3- 3.2 MB/s | 7.4 MB 00:02 2026-03-09T20:34:31.625 INFO:teuthology.orchestra.run.vm03.stdout:(24/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 2.4 MB/s | 5.4 MB 00:02 2026-03-09T20:34:31.652 INFO:teuthology.orchestra.run.vm08.stdout:(30/136): ceph-mgr-modules-core-19.2.3-678.ge91 2.1 MB/s | 253 kB 00:00 2026-03-09T20:34:31.654 INFO:teuthology.orchestra.run.vm04.stdout:(35/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 2.1 MB/s | 769 kB 00:00 2026-03-09T20:34:31.740 INFO:teuthology.orchestra.run.vm04.stdout:(36/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 475 kB/s | 40 kB 00:00 2026-03-09T20:34:31.746 INFO:teuthology.orchestra.run.vm03.stdout:(25/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.4 MB/s | 171 kB 00:00 2026-03-09T20:34:31.766 INFO:teuthology.orchestra.run.vm08.stdout:(31/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 432 kB/s | 49 kB 00:00 2026-03-09T20:34:31.797 INFO:teuthology.orchestra.run.vm04.stdout:(37/136): libconfig-1.7.2-9.el9.x86_64.rpm 1.2 MB/s | 72 kB 00:00 2026-03-09T20:34:31.809 INFO:teuthology.orchestra.run.vm04.stdout:(38/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 1.4 MB/s | 351 kB 00:00 2026-03-09T20:34:31.841 INFO:teuthology.orchestra.run.vm04.stdout:(39/136): libquadmath-11.5.0-14.el9.x86_64.rpm 5.7 MB/s | 184 kB 00:00 2026-03-09T20:34:31.865 INFO:teuthology.orchestra.run.vm03.stdout:(26/136): ceph-grafana-dashboards-19.2.3-678.ge 262 kB/s | 31 kB 00:00 2026-03-09T20:34:31.870 INFO:teuthology.orchestra.run.vm04.stdout:(40/136): mailcap-2.1.49-5.el9.noarch.rpm 1.1 MB/s | 33 kB 00:00 2026-03-09T20:34:31.879 INFO:teuthology.orchestra.run.vm08.stdout:(32/136): ceph-prometheus-alerts-19.2.3-678.ge9 147 kB/s | 17 kB 00:00 2026-03-09T20:34:31.887 INFO:teuthology.orchestra.run.vm04.stdout:(41/136): libgfortran-11.5.0-14.el9.x86_64.rpm 8.6 MB/s | 794 kB 00:00 2026-03-09T20:34:31.900 INFO:teuthology.orchestra.run.vm04.stdout:(42/136): pciutils-3.7.0-7.el9.x86_64.rpm 3.1 MB/s | 93 kB 00:00 2026-03-09T20:34:31.919 INFO:teuthology.orchestra.run.vm04.stdout:(43/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 7.9 MB/s | 253 kB 00:00 2026-03-09T20:34:31.949 INFO:teuthology.orchestra.run.vm04.stdout:(44/136): python3-ply-3.11-14.el9.noarch.rpm 3.5 MB/s | 106 kB 00:00 2026-03-09T20:34:31.979 INFO:teuthology.orchestra.run.vm04.stdout:(45/136): python3-pycparser-2.20-6.el9.noarch.r 4.4 MB/s | 135 kB 00:00 2026-03-09T20:34:31.986 INFO:teuthology.orchestra.run.vm03.stdout:(27/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.2 MB/s | 150 kB 00:00 2026-03-09T20:34:31.990 INFO:teuthology.orchestra.run.vm04.stdout:(46/136): python3-cryptography-36.0.1-5.el9.x86 14 MB/s | 1.2 MB 00:00 2026-03-09T20:34:31.997 INFO:teuthology.orchestra.run.vm08.stdout:(33/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.5 MB/s | 299 kB 00:00 2026-03-09T20:34:32.010 INFO:teuthology.orchestra.run.vm04.stdout:(47/136): python3-requests-2.25.1-10.el9.noarch 4.0 MB/s | 126 kB 00:00 2026-03-09T20:34:32.022 INFO:teuthology.orchestra.run.vm04.stdout:(48/136): python3-urllib3-1.26.5-7.el9.noarch.r 6.6 MB/s | 218 kB 00:00 2026-03-09T20:34:32.041 INFO:teuthology.orchestra.run.vm04.stdout:(49/136): unzip-6.0-59.el9.x86_64.rpm 5.8 MB/s | 182 kB 00:00 2026-03-09T20:34:32.054 INFO:teuthology.orchestra.run.vm04.stdout:(50/136): zip-3.0-35.el9.x86_64.rpm 8.2 MB/s | 266 kB 00:00 2026-03-09T20:34:32.120 INFO:teuthology.orchestra.run.vm08.stdout:(34/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 6.1 MB/s | 769 kB 00:00 2026-03-09T20:34:32.146 INFO:teuthology.orchestra.run.vm04.stdout:(51/136): flexiblas-3.0.4-9.el9.x86_64.rpm 325 kB/s | 30 kB 00:00 2026-03-09T20:34:32.206 INFO:teuthology.orchestra.run.vm04.stdout:(52/136): boost-program-options-1.75.0-13.el9.x 633 kB/s | 104 kB 00:00 2026-03-09T20:34:32.237 INFO:teuthology.orchestra.run.vm04.stdout:(53/136): flexiblas-openblas-openmp-3.0.4-9.el9 479 kB/s | 15 kB 00:00 2026-03-09T20:34:32.289 INFO:teuthology.orchestra.run.vm08.stdout:(35/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 2.0 MB/s | 351 kB 00:00 2026-03-09T20:34:32.318 INFO:teuthology.orchestra.run.vm08.stdout:(36/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 1.4 MB/s | 40 kB 00:00 2026-03-09T20:34:32.320 INFO:teuthology.orchestra.run.vm04.stdout:(54/136): libnbd-1.20.3-4.el9.x86_64.rpm 1.9 MB/s | 164 kB 00:00 2026-03-09T20:34:32.348 INFO:teuthology.orchestra.run.vm08.stdout:(37/136): libconfig-1.7.2-9.el9.x86_64.rpm 2.4 MB/s | 72 kB 00:00 2026-03-09T20:34:32.353 INFO:teuthology.orchestra.run.vm04.stdout:(55/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 4.8 MB/s | 160 kB 00:00 2026-03-09T20:34:32.370 INFO:teuthology.orchestra.run.vm04.stdout:(56/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 13 MB/s | 3.0 MB 00:00 2026-03-09T20:34:32.385 INFO:teuthology.orchestra.run.vm04.stdout:(57/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.4 MB/s | 45 kB 00:00 2026-03-09T20:34:32.406 INFO:teuthology.orchestra.run.vm08.stdout:(38/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 3.7 MB/s | 3.8 MB 00:01 2026-03-09T20:34:32.408 INFO:teuthology.orchestra.run.vm04.stdout:(58/136): librdkafka-1.6.1-102.el9.x86_64.rpm 17 MB/s | 662 kB 00:00 2026-03-09T20:34:32.413 INFO:teuthology.orchestra.run.vm08.stdout:(39/136): libgfortran-11.5.0-14.el9.x86_64.rpm 12 MB/s | 794 kB 00:00 2026-03-09T20:34:32.419 INFO:teuthology.orchestra.run.vm04.stdout:(59/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 7.3 MB/s | 246 kB 00:00 2026-03-09T20:34:32.442 INFO:teuthology.orchestra.run.vm04.stdout:(60/136): libxslt-1.1.34-12.el9.x86_64.rpm 6.9 MB/s | 233 kB 00:00 2026-03-09T20:34:32.442 INFO:teuthology.orchestra.run.vm08.stdout:(40/136): mailcap-2.1.49-5.el9.noarch.rpm 1.1 MB/s | 33 kB 00:00 2026-03-09T20:34:32.454 INFO:teuthology.orchestra.run.vm04.stdout:(61/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 8.3 MB/s | 292 kB 00:00 2026-03-09T20:34:32.472 INFO:teuthology.orchestra.run.vm08.stdout:(41/136): pciutils-3.7.0-7.el9.x86_64.rpm 3.0 MB/s | 93 kB 00:00 2026-03-09T20:34:32.474 INFO:teuthology.orchestra.run.vm04.stdout:(62/136): lua-5.4.4-4.el9.x86_64.rpm 5.7 MB/s | 188 kB 00:00 2026-03-09T20:34:32.485 INFO:teuthology.orchestra.run.vm04.stdout:(63/136): openblas-0.3.29-1.el9.x86_64.rpm 1.3 MB/s | 42 kB 00:00 2026-03-09T20:34:32.504 INFO:teuthology.orchestra.run.vm08.stdout:(42/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 7.9 MB/s | 253 kB 00:00 2026-03-09T20:34:32.550 INFO:teuthology.orchestra.run.vm08.stdout:(43/136): libquadmath-11.5.0-14.el9.x86_64.rpm 1.3 MB/s | 184 kB 00:00 2026-03-09T20:34:32.582 INFO:teuthology.orchestra.run.vm08.stdout:(44/136): python3-ply-3.11-14.el9.noarch.rpm 3.2 MB/s | 106 kB 00:00 2026-03-09T20:34:32.614 INFO:teuthology.orchestra.run.vm03.stdout:(28/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 2.6 MB/s | 3.1 MB 00:01 2026-03-09T20:34:32.614 INFO:teuthology.orchestra.run.vm08.stdout:(45/136): python3-pycparser-2.20-6.el9.noarch.r 4.2 MB/s | 135 kB 00:00 2026-03-09T20:34:32.618 INFO:teuthology.orchestra.run.vm04.stdout:(64/136): protobuf-3.14.0-17.el9.x86_64.rpm 7.6 MB/s | 1.0 MB 00:00 2026-03-09T20:34:32.645 INFO:teuthology.orchestra.run.vm08.stdout:(46/136): python3-requests-2.25.1-10.el9.noarch 4.1 MB/s | 126 kB 00:00 2026-03-09T20:34:32.650 INFO:teuthology.orchestra.run.vm08.stdout:(47/136): python3-cryptography-36.0.1-5.el9.x86 8.5 MB/s | 1.2 MB 00:00 2026-03-09T20:34:32.676 INFO:teuthology.orchestra.run.vm08.stdout:(48/136): python3-urllib3-1.26.5-7.el9.noarch.r 6.8 MB/s | 218 kB 00:00 2026-03-09T20:34:32.682 INFO:teuthology.orchestra.run.vm08.stdout:(49/136): unzip-6.0-59.el9.x86_64.rpm 5.7 MB/s | 182 kB 00:00 2026-03-09T20:34:32.709 INFO:teuthology.orchestra.run.vm08.stdout:(50/136): zip-3.0-35.el9.x86_64.rpm 8.1 MB/s | 266 kB 00:00 2026-03-09T20:34:32.741 INFO:teuthology.orchestra.run.vm04.stdout:(65/136): openblas-openmp-0.3.29-1.el9.x86_64.r 20 MB/s | 5.3 MB 00:00 2026-03-09T20:34:32.776 INFO:teuthology.orchestra.run.vm04.stdout:(66/136): python3-devel-3.9.25-3.el9.x86_64.rpm 7.0 MB/s | 244 kB 00:00 2026-03-09T20:34:32.812 INFO:teuthology.orchestra.run.vm04.stdout:(67/136): python3-jinja2-2.11.3-8.el9.noarch.rp 6.8 MB/s | 249 kB 00:00 2026-03-09T20:34:32.844 INFO:teuthology.orchestra.run.vm04.stdout:(68/136): python3-jmespath-1.0.1-1.el9.noarch.r 1.5 MB/s | 48 kB 00:00 2026-03-09T20:34:32.877 INFO:teuthology.orchestra.run.vm04.stdout:(69/136): python3-libstoragemgmt-1.10.1-1.el9.x 5.3 MB/s | 177 kB 00:00 2026-03-09T20:34:32.906 INFO:teuthology.orchestra.run.vm08.stdout:(51/136): flexiblas-3.0.4-9.el9.x86_64.rpm 150 kB/s | 30 kB 00:00 2026-03-09T20:34:32.909 INFO:teuthology.orchestra.run.vm04.stdout:(70/136): python3-mako-1.1.4-6.el9.noarch.rpm 5.2 MB/s | 172 kB 00:00 2026-03-09T20:34:32.940 INFO:teuthology.orchestra.run.vm04.stdout:(71/136): python3-markupsafe-1.1.1-12.el9.x86_6 1.1 MB/s | 35 kB 00:00 2026-03-09T20:34:32.978 INFO:teuthology.orchestra.run.vm08.stdout:(52/136): boost-program-options-1.75.0-13.el9.x 352 kB/s | 104 kB 00:00 2026-03-09T20:34:33.075 INFO:teuthology.orchestra.run.vm04.stdout:(72/136): python3-babel-2.9.1-2.el9.noarch.rpm 13 MB/s | 6.0 MB 00:00 2026-03-09T20:34:33.075 INFO:teuthology.orchestra.run.vm08.stdout:(53/136): flexiblas-openblas-openmp-3.0.4-9.el9 153 kB/s | 15 kB 00:00 2026-03-09T20:34:33.223 INFO:teuthology.orchestra.run.vm08.stdout:(54/136): libnbd-1.20.3-4.el9.x86_64.rpm 1.1 MB/s | 164 kB 00:00 2026-03-09T20:34:33.277 INFO:teuthology.orchestra.run.vm04.stdout:(73/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 18 MB/s | 6.1 MB 00:00 2026-03-09T20:34:33.311 INFO:teuthology.orchestra.run.vm04.stdout:(74/136): python3-packaging-20.9-5.el9.noarch.r 2.2 MB/s | 77 kB 00:00 2026-03-09T20:34:33.320 INFO:teuthology.orchestra.run.vm08.stdout:(55/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 1.6 MB/s | 160 kB 00:00 2026-03-09T20:34:33.358 INFO:teuthology.orchestra.run.vm04.stdout:(75/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 1.5 MB/s | 442 kB 00:00 2026-03-09T20:34:33.391 INFO:teuthology.orchestra.run.vm04.stdout:(76/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 4.6 MB/s | 157 kB 00:00 2026-03-09T20:34:33.399 INFO:teuthology.orchestra.run.vm08.stdout:(56/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 578 kB/s | 45 kB 00:00 2026-03-09T20:34:33.399 INFO:teuthology.orchestra.run.vm04.stdout:(77/136): python3-protobuf-3.14.0-17.el9.noarch 3.0 MB/s | 267 kB 00:00 2026-03-09T20:34:33.426 INFO:teuthology.orchestra.run.vm04.stdout:(78/136): python3-pyasn1-modules-0.4.8-7.el9.no 7.9 MB/s | 277 kB 00:00 2026-03-09T20:34:33.426 INFO:teuthology.orchestra.run.vm03.stdout:(29/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 2.6 MB/s | 3.8 MB 00:01 2026-03-09T20:34:33.430 INFO:teuthology.orchestra.run.vm04.stdout:(79/136): python3-requests-oauthlib-1.3.0-12.el 1.7 MB/s | 54 kB 00:00 2026-03-09T20:34:33.461 INFO:teuthology.orchestra.run.vm04.stdout:(80/136): python3-toml-0.10.2-6.el9.noarch.rpm 1.3 MB/s | 42 kB 00:00 2026-03-09T20:34:33.514 INFO:teuthology.orchestra.run.vm04.stdout:(81/136): qatlib-25.08.0-2.el9.x86_64.rpm 4.5 MB/s | 240 kB 00:00 2026-03-09T20:34:33.545 INFO:teuthology.orchestra.run.vm08.stdout:(57/136): librdkafka-1.6.1-102.el9.x86_64.rpm 4.5 MB/s | 662 kB 00:00 2026-03-09T20:34:33.550 INFO:teuthology.orchestra.run.vm03.stdout:(30/136): ceph-mgr-modules-core-19.2.3-678.ge91 2.0 MB/s | 253 kB 00:00 2026-03-09T20:34:33.560 INFO:teuthology.orchestra.run.vm04.stdout:(82/136): qatlib-service-25.08.0-2.el9.x86_64.r 799 kB/s | 37 kB 00:00 2026-03-09T20:34:33.609 INFO:teuthology.orchestra.run.vm04.stdout:(83/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 1.3 MB/s | 66 kB 00:00 2026-03-09T20:34:33.644 INFO:teuthology.orchestra.run.vm04.stdout:(84/136): socat-1.7.4.1-8.el9.x86_64.rpm 8.5 MB/s | 303 kB 00:00 2026-03-09T20:34:33.647 INFO:teuthology.orchestra.run.vm08.stdout:(58/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 2.4 MB/s | 246 kB 00:00 2026-03-09T20:34:33.670 INFO:teuthology.orchestra.run.vm03.stdout:(31/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 413 kB/s | 49 kB 00:00 2026-03-09T20:34:33.676 INFO:teuthology.orchestra.run.vm04.stdout:(85/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 2.0 MB/s | 64 kB 00:00 2026-03-09T20:34:33.734 INFO:teuthology.orchestra.run.vm08.stdout:(59/136): libxslt-1.1.34-12.el9.x86_64.rpm 2.6 MB/s | 233 kB 00:00 2026-03-09T20:34:33.771 INFO:teuthology.orchestra.run.vm04.stdout:(86/136): lua-devel-5.4.4-4.el9.x86_64.rpm 236 kB/s | 22 kB 00:00 2026-03-09T20:34:33.783 INFO:teuthology.orchestra.run.vm08.stdout:(60/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 3.4 MB/s | 3.0 MB 00:00 2026-03-09T20:34:33.790 INFO:teuthology.orchestra.run.vm03.stdout:(32/136): ceph-prometheus-alerts-19.2.3-678.ge9 139 kB/s | 17 kB 00:00 2026-03-09T20:34:33.824 INFO:teuthology.orchestra.run.vm08.stdout:(61/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 3.2 MB/s | 292 kB 00:00 2026-03-09T20:34:33.884 INFO:teuthology.orchestra.run.vm08.stdout:(62/136): lua-5.4.4-4.el9.x86_64.rpm 1.8 MB/s | 188 kB 00:00 2026-03-09T20:34:33.909 INFO:teuthology.orchestra.run.vm08.stdout:(63/136): openblas-0.3.29-1.el9.x86_64.rpm 500 kB/s | 42 kB 00:00 2026-03-09T20:34:33.913 INFO:teuthology.orchestra.run.vm03.stdout:(33/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 299 kB 00:00 2026-03-09T20:34:33.925 INFO:teuthology.orchestra.run.vm04.stdout:(87/136): protobuf-compiler-3.14.0-17.el9.x86_6 5.5 MB/s | 862 kB 00:00 2026-03-09T20:34:33.942 INFO:teuthology.orchestra.run.vm04.stdout:(88/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 33 MB/s | 551 kB 00:00 2026-03-09T20:34:33.949 INFO:teuthology.orchestra.run.vm04.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 43 MB/s | 308 kB 00:00 2026-03-09T20:34:33.951 INFO:teuthology.orchestra.run.vm04.stdout:(90/136): grpc-data-1.46.7-10.el9.noarch.rpm 8.1 MB/s | 19 kB 00:00 2026-03-09T20:34:34.013 INFO:teuthology.orchestra.run.vm04.stdout:(91/136): libarrow-9.0.0-15.el9.x86_64.rpm 73 MB/s | 4.4 MB 00:00 2026-03-09T20:34:34.016 INFO:teuthology.orchestra.run.vm04.stdout:(92/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 9.2 MB/s | 25 kB 00:00 2026-03-09T20:34:34.019 INFO:teuthology.orchestra.run.vm04.stdout:(93/136): liboath-2.6.12-1.el9.x86_64.rpm 18 MB/s | 49 kB 00:00 2026-03-09T20:34:34.022 INFO:teuthology.orchestra.run.vm04.stdout:(94/136): libunwind-1.6.2-1.el9.x86_64.rpm 21 MB/s | 67 kB 00:00 2026-03-09T20:34:34.030 INFO:teuthology.orchestra.run.vm04.stdout:(95/136): luarocks-3.9.2-5.el9.noarch.rpm 20 MB/s | 151 kB 00:00 2026-03-09T20:34:34.045 INFO:teuthology.orchestra.run.vm04.stdout:(96/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 57 MB/s | 838 kB 00:00 2026-03-09T20:34:34.057 INFO:teuthology.orchestra.run.vm04.stdout:(97/136): python3-asyncssh-2.13.2-5.el9.noarch. 45 MB/s | 548 kB 00:00 2026-03-09T20:34:34.060 INFO:teuthology.orchestra.run.vm04.stdout:(98/136): python3-autocommand-2.2.2-8.el9.noarc 9.4 MB/s | 29 kB 00:00 2026-03-09T20:34:34.063 INFO:teuthology.orchestra.run.vm08.stdout:(64/136): protobuf-3.14.0-17.el9.x86_64.rpm 6.5 MB/s | 1.0 MB 00:00 2026-03-09T20:34:34.063 INFO:teuthology.orchestra.run.vm04.stdout:(99/136): python3-backports-tarfile-1.2.0-1.el9 22 MB/s | 60 kB 00:00 2026-03-09T20:34:34.066 INFO:teuthology.orchestra.run.vm04.stdout:(100/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 15 MB/s | 43 kB 00:00 2026-03-09T20:34:34.069 INFO:teuthology.orchestra.run.vm04.stdout:(101/136): python3-cachetools-4.2.4-1.el9.noarc 11 MB/s | 32 kB 00:00 2026-03-09T20:34:34.072 INFO:teuthology.orchestra.run.vm04.stdout:(102/136): python3-certifi-2023.05.07-4.el9.noa 6.0 MB/s | 14 kB 00:00 2026-03-09T20:34:34.077 INFO:teuthology.orchestra.run.vm04.stdout:(103/136): python3-cheroot-10.0.1-4.el9.noarch. 34 MB/s | 173 kB 00:00 2026-03-09T20:34:34.084 INFO:teuthology.orchestra.run.vm04.stdout:(104/136): python3-cherrypy-18.6.1-2.el9.noarch 54 MB/s | 358 kB 00:00 2026-03-09T20:34:34.090 INFO:teuthology.orchestra.run.vm04.stdout:(105/136): python3-google-auth-2.45.0-1.el9.noa 47 MB/s | 254 kB 00:00 2026-03-09T20:34:34.117 INFO:teuthology.orchestra.run.vm04.stdout:(106/136): python3-grpcio-1.46.7-10.el9.x86_64. 76 MB/s | 2.0 MB 00:00 2026-03-09T20:34:34.123 INFO:teuthology.orchestra.run.vm04.stdout:(107/136): python3-grpcio-tools-1.46.7-10.el9.x 27 MB/s | 144 kB 00:00 2026-03-09T20:34:34.125 INFO:teuthology.orchestra.run.vm04.stdout:(108/136): python3-jaraco-8.2.1-3.el9.noarch.rp 5.1 MB/s | 11 kB 00:00 2026-03-09T20:34:34.128 INFO:teuthology.orchestra.run.vm04.stdout:(109/136): python3-jaraco-classes-3.2.1-5.el9.n 8.0 MB/s | 18 kB 00:00 2026-03-09T20:34:34.130 INFO:teuthology.orchestra.run.vm04.stdout:(110/136): python3-jaraco-collections-3.0.0-8.e 9.4 MB/s | 23 kB 00:00 2026-03-09T20:34:34.132 INFO:teuthology.orchestra.run.vm04.stdout:(111/136): python3-jaraco-context-6.0.1-3.el9.n 8.7 MB/s | 20 kB 00:00 2026-03-09T20:34:34.135 INFO:teuthology.orchestra.run.vm04.stdout:(112/136): python3-jaraco-functools-3.5.0-2.el9 8.7 MB/s | 19 kB 00:00 2026-03-09T20:34:34.137 INFO:teuthology.orchestra.run.vm04.stdout:(113/136): python3-jaraco-text-4.0.0-2.el9.noar 11 MB/s | 26 kB 00:00 2026-03-09T20:34:34.159 INFO:teuthology.orchestra.run.vm04.stdout:(114/136): python3-kubernetes-26.1.0-3.el9.noar 47 MB/s | 1.0 MB 00:00 2026-03-09T20:34:34.164 INFO:teuthology.orchestra.run.vm04.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 11 MB/s | 46 kB 00:00 2026-03-09T20:34:34.168 INFO:teuthology.orchestra.run.vm04.stdout:(116/136): python3-more-itertools-8.12.0-2.el9. 21 MB/s | 79 kB 00:00 2026-03-09T20:34:34.173 INFO:teuthology.orchestra.run.vm04.stdout:(117/136): python3-natsort-7.1.1-5.el9.noarch.r 11 MB/s | 58 kB 00:00 2026-03-09T20:34:34.181 INFO:teuthology.orchestra.run.vm04.stdout:(118/136): python3-pecan-1.4.2-3.el9.noarch.rpm 37 MB/s | 272 kB 00:00 2026-03-09T20:34:34.195 INFO:teuthology.orchestra.run.vm04.stdout:(119/136): python3-portend-3.1.0-2.el9.noarch.r 1.2 MB/s | 16 kB 00:00 2026-03-09T20:34:34.199 INFO:teuthology.orchestra.run.vm04.stdout:(120/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 23 MB/s | 90 kB 00:00 2026-03-09T20:34:34.201 INFO:teuthology.orchestra.run.vm04.stdout:(121/136): python3-repoze-lru-0.7-16.el9.noarch 12 MB/s | 31 kB 00:00 2026-03-09T20:34:34.207 INFO:teuthology.orchestra.run.vm04.stdout:(122/136): python3-routes-2.5.1-5.el9.noarch.rp 38 MB/s | 188 kB 00:00 2026-03-09T20:34:34.210 INFO:teuthology.orchestra.run.vm04.stdout:(123/136): python3-rsa-4.9-2.el9.noarch.rpm 21 MB/s | 59 kB 00:00 2026-03-09T20:34:34.212 INFO:teuthology.orchestra.run.vm04.stdout:(124/136): python3-tempora-5.0.0-2.el9.noarch.r 14 MB/s | 36 kB 00:00 2026-03-09T20:34:34.216 INFO:teuthology.orchestra.run.vm04.stdout:(125/136): python3-typing-extensions-4.15.0-1.e 25 MB/s | 86 kB 00:00 2026-03-09T20:34:34.224 INFO:teuthology.orchestra.run.vm04.stdout:(126/136): python3-webob-1.8.8-2.el9.noarch.rpm 29 MB/s | 230 kB 00:00 2026-03-09T20:34:34.230 INFO:teuthology.orchestra.run.vm04.stdout:(127/136): python3-websocket-client-1.2.3-2.el9 16 MB/s | 90 kB 00:00 2026-03-09T20:34:34.243 INFO:teuthology.orchestra.run.vm04.stdout:(128/136): python3-werkzeug-2.0.3-3.el9.1.noarc 33 MB/s | 427 kB 00:00 2026-03-09T20:34:34.251 INFO:teuthology.orchestra.run.vm04.stdout:(129/136): python3-xmltodict-0.12.0-15.el9.noar 3.0 MB/s | 22 kB 00:00 2026-03-09T20:34:34.253 INFO:teuthology.orchestra.run.vm04.stdout:(130/136): python3-zc-lockfile-2.0-10.el9.noarc 8.9 MB/s | 20 kB 00:00 2026-03-09T20:34:34.259 INFO:teuthology.orchestra.run.vm04.stdout:(131/136): re2-20211101-20.el9.x86_64.rpm 38 MB/s | 191 kB 00:00 2026-03-09T20:34:34.277 INFO:teuthology.orchestra.run.vm03.stdout:(34/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 2.1 MB/s | 769 kB 00:00 2026-03-09T20:34:34.293 INFO:teuthology.orchestra.run.vm04.stdout:(132/136): thrift-0.15.0-4.el9.x86_64.rpm 47 MB/s | 1.6 MB 00:00 2026-03-09T20:34:34.456 INFO:teuthology.orchestra.run.vm03.stdout:(35/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 1.9 MB/s | 351 kB 00:00 2026-03-09T20:34:34.488 INFO:teuthology.orchestra.run.vm03.stdout:(36/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 1.3 MB/s | 40 kB 00:00 2026-03-09T20:34:34.536 INFO:teuthology.orchestra.run.vm03.stdout:(37/136): libconfig-1.7.2-9.el9.x86_64.rpm 1.5 MB/s | 72 kB 00:00 2026-03-09T20:34:34.614 INFO:teuthology.orchestra.run.vm03.stdout:(38/136): libgfortran-11.5.0-14.el9.x86_64.rpm 9.9 MB/s | 794 kB 00:00 2026-03-09T20:34:34.665 INFO:teuthology.orchestra.run.vm03.stdout:(39/136): libquadmath-11.5.0-14.el9.x86_64.rpm 3.6 MB/s | 184 kB 00:00 2026-03-09T20:34:34.696 INFO:teuthology.orchestra.run.vm03.stdout:(40/136): mailcap-2.1.49-5.el9.noarch.rpm 1.1 MB/s | 33 kB 00:00 2026-03-09T20:34:34.729 INFO:teuthology.orchestra.run.vm03.stdout:(41/136): pciutils-3.7.0-7.el9.x86_64.rpm 2.8 MB/s | 93 kB 00:00 2026-03-09T20:34:34.777 INFO:teuthology.orchestra.run.vm03.stdout:(42/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 5.3 MB/s | 253 kB 00:00 2026-03-09T20:34:34.869 INFO:teuthology.orchestra.run.vm03.stdout:(43/136): python3-cryptography-36.0.1-5.el9.x86 13 MB/s | 1.2 MB 00:00 2026-03-09T20:34:34.918 INFO:teuthology.orchestra.run.vm03.stdout:(44/136): python3-ply-3.11-14.el9.noarch.rpm 2.2 MB/s | 106 kB 00:00 2026-03-09T20:34:34.958 INFO:teuthology.orchestra.run.vm03.stdout:(45/136): python3-pycparser-2.20-6.el9.noarch.r 3.4 MB/s | 135 kB 00:00 2026-03-09T20:34:34.995 INFO:teuthology.orchestra.run.vm03.stdout:(46/136): ceph-mgr-diskprediction-local-19.2.3- 3.1 MB/s | 7.4 MB 00:02 2026-03-09T20:34:35.011 INFO:teuthology.orchestra.run.vm03.stdout:(47/136): python3-requests-2.25.1-10.el9.noarch 2.3 MB/s | 126 kB 00:00 2026-03-09T20:34:35.044 INFO:teuthology.orchestra.run.vm03.stdout:(48/136): unzip-6.0-59.el9.x86_64.rpm 5.4 MB/s | 182 kB 00:00 2026-03-09T20:34:35.093 INFO:teuthology.orchestra.run.vm03.stdout:(49/136): zip-3.0-35.el9.x86_64.rpm 5.3 MB/s | 266 kB 00:00 2026-03-09T20:34:35.108 INFO:teuthology.orchestra.run.vm08.stdout:(65/136): openblas-openmp-0.3.29-1.el9.x86_64.r 4.3 MB/s | 5.3 MB 00:01 2026-03-09T20:34:35.172 INFO:teuthology.orchestra.run.vm03.stdout:(50/136): python3-urllib3-1.26.5-7.el9.noarch.r 1.2 MB/s | 218 kB 00:00 2026-03-09T20:34:35.264 INFO:teuthology.orchestra.run.vm03.stdout:(51/136): boost-program-options-1.75.0-13.el9.x 611 kB/s | 104 kB 00:00 2026-03-09T20:34:35.281 INFO:teuthology.orchestra.run.vm03.stdout:(52/136): flexiblas-3.0.4-9.el9.x86_64.rpm 272 kB/s | 30 kB 00:00 2026-03-09T20:34:35.281 INFO:teuthology.orchestra.run.vm08.stdout:(66/136): ceph-mgr-diskprediction-local-19.2.3- 2.0 MB/s | 7.4 MB 00:03 2026-03-09T20:34:35.283 INFO:teuthology.orchestra.run.vm08.stdout:(67/136): python3-devel-3.9.25-3.el9.x86_64.rpm 1.4 MB/s | 244 kB 00:00 2026-03-09T20:34:35.334 INFO:teuthology.orchestra.run.vm03.stdout:(53/136): flexiblas-openblas-openmp-3.0.4-9.el9 280 kB/s | 15 kB 00:00 2026-03-09T20:34:35.363 INFO:teuthology.orchestra.run.vm08.stdout:(68/136): python3-jmespath-1.0.1-1.el9.noarch.r 602 kB/s | 48 kB 00:00 2026-03-09T20:34:35.475 INFO:teuthology.orchestra.run.vm08.stdout:(69/136): python3-babel-2.9.1-2.el9.noarch.rpm 4.2 MB/s | 6.0 MB 00:01 2026-03-09T20:34:35.485 INFO:teuthology.orchestra.run.vm03.stdout:(54/136): libnbd-1.20.3-4.el9.x86_64.rpm 1.1 MB/s | 164 kB 00:00 2026-03-09T20:34:35.496 INFO:teuthology.orchestra.run.vm08.stdout:(70/136): python3-libstoragemgmt-1.10.1-1.el9.x 1.3 MB/s | 177 kB 00:00 2026-03-09T20:34:35.532 INFO:teuthology.orchestra.run.vm04.stdout:(133/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 9.1 MB/s | 19 MB 00:02 2026-03-09T20:34:35.554 INFO:teuthology.orchestra.run.vm03.stdout:(55/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 2.3 MB/s | 160 kB 00:00 2026-03-09T20:34:35.590 INFO:teuthology.orchestra.run.vm08.stdout:(71/136): python3-markupsafe-1.1.1-12.el9.x86_6 371 kB/s | 35 kB 00:00 2026-03-09T20:34:35.602 INFO:teuthology.orchestra.run.vm03.stdout:(56/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 960 kB/s | 45 kB 00:00 2026-03-09T20:34:35.621 INFO:teuthology.orchestra.run.vm08.stdout:(72/136): python3-mako-1.1.4-6.el9.noarch.rpm 1.2 MB/s | 172 kB 00:00 2026-03-09T20:34:35.646 INFO:teuthology.orchestra.run.vm08.stdout:(73/136): python3-jinja2-2.11.3-8.el9.noarch.rp 684 kB/s | 249 kB 00:00 2026-03-09T20:34:35.666 INFO:teuthology.orchestra.run.vm03.stdout:(57/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 7.4 MB/s | 3.0 MB 00:00 2026-03-09T20:34:35.696 INFO:teuthology.orchestra.run.vm03.stdout:(58/136): librdkafka-1.6.1-102.el9.x86_64.rpm 6.9 MB/s | 662 kB 00:00 2026-03-09T20:34:35.724 INFO:teuthology.orchestra.run.vm03.stdout:(59/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 4.2 MB/s | 246 kB 00:00 2026-03-09T20:34:35.740 INFO:teuthology.orchestra.run.vm08.stdout:(74/136): python3-packaging-20.9-5.el9.noarch.r 821 kB/s | 77 kB 00:00 2026-03-09T20:34:35.758 INFO:teuthology.orchestra.run.vm03.stdout:(60/136): libxslt-1.1.34-12.el9.x86_64.rpm 3.7 MB/s | 233 kB 00:00 2026-03-09T20:34:35.805 INFO:teuthology.orchestra.run.vm04.stdout:(134/136): librados2-19.2.3-678.ge911bdeb.el9.x 2.3 MB/s | 3.4 MB 00:01 2026-03-09T20:34:35.813 INFO:teuthology.orchestra.run.vm03.stdout:(61/136): lua-5.4.4-4.el9.x86_64.rpm 3.4 MB/s | 188 kB 00:00 2026-03-09T20:34:35.816 INFO:teuthology.orchestra.run.vm03.stdout:(62/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 3.1 MB/s | 292 kB 00:00 2026-03-09T20:34:35.859 INFO:teuthology.orchestra.run.vm03.stdout:(63/136): openblas-0.3.29-1.el9.x86_64.rpm 924 kB/s | 42 kB 00:00 2026-03-09T20:34:35.878 INFO:teuthology.orchestra.run.vm08.stdout:(75/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 1.7 MB/s | 442 kB 00:00 2026-03-09T20:34:35.898 INFO:teuthology.orchestra.run.vm08.stdout:(76/136): python3-protobuf-3.14.0-17.el9.noarch 1.7 MB/s | 267 kB 00:00 2026-03-09T20:34:36.025 INFO:teuthology.orchestra.run.vm08.stdout:(77/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 1.0 MB/s | 157 kB 00:00 2026-03-09T20:34:36.036 INFO:teuthology.orchestra.run.vm03.stdout:(64/136): protobuf-3.14.0-17.el9.x86_64.rpm 5.7 MB/s | 1.0 MB 00:00 2026-03-09T20:34:36.125 INFO:teuthology.orchestra.run.vm08.stdout:(78/136): python3-requests-oauthlib-1.3.0-12.el 543 kB/s | 54 kB 00:00 2026-03-09T20:34:36.163 INFO:teuthology.orchestra.run.vm08.stdout:(79/136): python3-pyasn1-modules-0.4.8-7.el9.no 1.0 MB/s | 277 kB 00:00 2026-03-09T20:34:36.273 INFO:teuthology.orchestra.run.vm08.stdout:(80/136): python3-toml-0.10.2-6.el9.noarch.rpm 380 kB/s | 42 kB 00:00 2026-03-09T20:34:36.395 INFO:teuthology.orchestra.run.vm08.stdout:(81/136): qatlib-25.08.0-2.el9.x86_64.rpm 1.9 MB/s | 240 kB 00:00 2026-03-09T20:34:36.465 INFO:teuthology.orchestra.run.vm08.stdout:(82/136): qatlib-service-25.08.0-2.el9.x86_64.r 534 kB/s | 37 kB 00:00 2026-03-09T20:34:36.568 INFO:teuthology.orchestra.run.vm08.stdout:(83/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 642 kB/s | 66 kB 00:00 2026-03-09T20:34:36.681 INFO:teuthology.orchestra.run.vm08.stdout:(84/136): socat-1.7.4.1-8.el9.x86_64.rpm 2.6 MB/s | 303 kB 00:00 2026-03-09T20:34:36.689 INFO:teuthology.orchestra.run.vm03.stdout:(65/136): python3-babel-2.9.1-2.el9.noarch.rpm 9.1 MB/s | 6.0 MB 00:00 2026-03-09T20:34:36.762 INFO:teuthology.orchestra.run.vm08.stdout:(85/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 786 kB/s | 64 kB 00:00 2026-03-09T20:34:36.779 INFO:teuthology.orchestra.run.vm03.stdout:(66/136): python3-devel-3.9.25-3.el9.x86_64.rpm 2.7 MB/s | 244 kB 00:00 2026-03-09T20:34:36.834 INFO:teuthology.orchestra.run.vm03.stdout:(67/136): python3-jinja2-2.11.3-8.el9.noarch.rp 4.4 MB/s | 249 kB 00:00 2026-03-09T20:34:36.845 INFO:teuthology.orchestra.run.vm08.stdout:(86/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 4.9 MB/s | 6.1 MB 00:01 2026-03-09T20:34:36.884 INFO:teuthology.orchestra.run.vm03.stdout:(68/136): python3-jmespath-1.0.1-1.el9.noarch.r 950 kB/s | 48 kB 00:00 2026-03-09T20:34:36.909 INFO:teuthology.orchestra.run.vm08.stdout:(87/136): lua-devel-5.4.4-4.el9.x86_64.rpm 152 kB/s | 22 kB 00:00 2026-03-09T20:34:36.924 INFO:teuthology.orchestra.run.vm03.stdout:(69/136): openblas-openmp-0.3.29-1.el9.x86_64.r 4.8 MB/s | 5.3 MB 00:01 2026-03-09T20:34:36.927 INFO:teuthology.orchestra.run.vm08.stdout:(88/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 31 MB/s | 551 kB 00:00 2026-03-09T20:34:36.935 INFO:teuthology.orchestra.run.vm08.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 39 MB/s | 308 kB 00:00 2026-03-09T20:34:36.937 INFO:teuthology.orchestra.run.vm08.stdout:(90/136): grpc-data-1.46.7-10.el9.noarch.rpm 8.2 MB/s | 19 kB 00:00 2026-03-09T20:34:36.963 INFO:teuthology.orchestra.run.vm03.stdout:(70/136): python3-libstoragemgmt-1.10.1-1.el9.x 2.2 MB/s | 177 kB 00:00 2026-03-09T20:34:36.980 INFO:teuthology.orchestra.run.vm03.stdout:(71/136): python3-mako-1.1.4-6.el9.noarch.rpm 3.0 MB/s | 172 kB 00:00 2026-03-09T20:34:37.005 INFO:teuthology.orchestra.run.vm08.stdout:(91/136): libarrow-9.0.0-15.el9.x86_64.rpm 65 MB/s | 4.4 MB 00:00 2026-03-09T20:34:37.005 INFO:teuthology.orchestra.run.vm03.stdout:(72/136): python3-markupsafe-1.1.1-12.el9.x86_6 822 kB/s | 35 kB 00:00 2026-03-09T20:34:37.008 INFO:teuthology.orchestra.run.vm08.stdout:(92/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 7.7 MB/s | 25 kB 00:00 2026-03-09T20:34:37.011 INFO:teuthology.orchestra.run.vm08.stdout:(93/136): liboath-2.6.12-1.el9.x86_64.rpm 19 MB/s | 49 kB 00:00 2026-03-09T20:34:37.014 INFO:teuthology.orchestra.run.vm08.stdout:(94/136): libunwind-1.6.2-1.el9.x86_64.rpm 24 MB/s | 67 kB 00:00 2026-03-09T20:34:37.019 INFO:teuthology.orchestra.run.vm08.stdout:(95/136): luarocks-3.9.2-5.el9.noarch.rpm 35 MB/s | 151 kB 00:00 2026-03-09T20:34:37.030 INFO:teuthology.orchestra.run.vm08.stdout:(96/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 72 MB/s | 838 kB 00:00 2026-03-09T20:34:37.040 INFO:teuthology.orchestra.run.vm08.stdout:(97/136): python3-asyncssh-2.13.2-5.el9.noarch. 56 MB/s | 548 kB 00:00 2026-03-09T20:34:37.043 INFO:teuthology.orchestra.run.vm08.stdout:(98/136): python3-autocommand-2.2.2-8.el9.noarc 13 MB/s | 29 kB 00:00 2026-03-09T20:34:37.046 INFO:teuthology.orchestra.run.vm08.stdout:(99/136): python3-backports-tarfile-1.2.0-1.el9 21 MB/s | 60 kB 00:00 2026-03-09T20:34:37.058 INFO:teuthology.orchestra.run.vm08.stdout:(100/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 3.9 MB/s | 43 kB 00:00 2026-03-09T20:34:37.061 INFO:teuthology.orchestra.run.vm08.stdout:(101/136): python3-cachetools-4.2.4-1.el9.noarc 8.8 MB/s | 32 kB 00:00 2026-03-09T20:34:37.064 INFO:teuthology.orchestra.run.vm08.stdout:(102/136): python3-certifi-2023.05.07-4.el9.noa 6.4 MB/s | 14 kB 00:00 2026-03-09T20:34:37.069 INFO:teuthology.orchestra.run.vm08.stdout:(103/136): python3-cheroot-10.0.1-4.el9.noarch. 34 MB/s | 173 kB 00:00 2026-03-09T20:34:37.085 INFO:teuthology.orchestra.run.vm08.stdout:(104/136): python3-cherrypy-18.6.1-2.el9.noarch 23 MB/s | 358 kB 00:00 2026-03-09T20:34:37.090 INFO:teuthology.orchestra.run.vm08.stdout:(105/136): python3-google-auth-2.45.0-1.el9.noa 49 MB/s | 254 kB 00:00 2026-03-09T20:34:37.117 INFO:teuthology.orchestra.run.vm08.stdout:(106/136): python3-grpcio-1.46.7-10.el9.x86_64. 77 MB/s | 2.0 MB 00:00 2026-03-09T20:34:37.122 INFO:teuthology.orchestra.run.vm08.stdout:(107/136): python3-grpcio-tools-1.46.7-10.el9.x 33 MB/s | 144 kB 00:00 2026-03-09T20:34:37.124 INFO:teuthology.orchestra.run.vm08.stdout:(108/136): python3-jaraco-8.2.1-3.el9.noarch.rp 4.7 MB/s | 11 kB 00:00 2026-03-09T20:34:37.127 INFO:teuthology.orchestra.run.vm08.stdout:(109/136): python3-jaraco-classes-3.2.1-5.el9.n 6.2 MB/s | 18 kB 00:00 2026-03-09T20:34:37.129 INFO:teuthology.orchestra.run.vm08.stdout:(110/136): python3-jaraco-collections-3.0.0-8.e 11 MB/s | 23 kB 00:00 2026-03-09T20:34:37.132 INFO:teuthology.orchestra.run.vm08.stdout:(111/136): python3-jaraco-context-6.0.1-3.el9.n 9.0 MB/s | 20 kB 00:00 2026-03-09T20:34:37.134 INFO:teuthology.orchestra.run.vm08.stdout:(112/136): python3-jaraco-functools-3.5.0-2.el9 8.8 MB/s | 19 kB 00:00 2026-03-09T20:34:37.136 INFO:teuthology.orchestra.run.vm08.stdout:(113/136): python3-jaraco-text-4.0.0-2.el9.noar 12 MB/s | 26 kB 00:00 2026-03-09T20:34:37.151 INFO:teuthology.orchestra.run.vm08.stdout:(114/136): python3-kubernetes-26.1.0-3.el9.noar 68 MB/s | 1.0 MB 00:00 2026-03-09T20:34:37.156 INFO:teuthology.orchestra.run.vm08.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 12 MB/s | 46 kB 00:00 2026-03-09T20:34:37.158 INFO:teuthology.orchestra.run.vm08.stdout:(116/136): python3-more-itertools-8.12.0-2.el9. 28 MB/s | 79 kB 00:00 2026-03-09T20:34:37.162 INFO:teuthology.orchestra.run.vm08.stdout:(117/136): python3-natsort-7.1.1-5.el9.noarch.r 14 MB/s | 58 kB 00:00 2026-03-09T20:34:37.167 INFO:teuthology.orchestra.run.vm08.stdout:(118/136): python3-pecan-1.4.2-3.el9.noarch.rpm 54 MB/s | 272 kB 00:00 2026-03-09T20:34:37.171 INFO:teuthology.orchestra.run.vm08.stdout:(119/136): python3-portend-3.1.0-2.el9.noarch.r 4.2 MB/s | 16 kB 00:00 2026-03-09T20:34:37.177 INFO:teuthology.orchestra.run.vm08.stdout:(120/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 15 MB/s | 90 kB 00:00 2026-03-09T20:34:37.180 INFO:teuthology.orchestra.run.vm08.stdout:(121/136): protobuf-compiler-3.14.0-17.el9.x86_ 2.5 MB/s | 862 kB 00:00 2026-03-09T20:34:37.189 INFO:teuthology.orchestra.run.vm08.stdout:(122/136): python3-routes-2.5.1-5.el9.noarch.rp 23 MB/s | 188 kB 00:00 2026-03-09T20:34:37.189 INFO:teuthology.orchestra.run.vm04.stdout:(135/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 1.9 MB/s | 3.2 MB 00:01 2026-03-09T20:34:37.189 INFO:teuthology.orchestra.run.vm08.stdout:(123/136): python3-repoze-lru-0.7-16.el9.noarch 2.6 MB/s | 31 kB 00:00 2026-03-09T20:34:37.192 INFO:teuthology.orchestra.run.vm08.stdout:(124/136): python3-rsa-4.9-2.el9.noarch.rpm 17 MB/s | 59 kB 00:00 2026-03-09T20:34:37.193 INFO:teuthology.orchestra.run.vm08.stdout:(125/136): python3-tempora-5.0.0-2.el9.noarch.r 11 MB/s | 36 kB 00:00 2026-03-09T20:34:37.196 INFO:teuthology.orchestra.run.vm08.stdout:(126/136): python3-typing-extensions-4.15.0-1.e 21 MB/s | 86 kB 00:00 2026-03-09T20:34:37.198 INFO:teuthology.orchestra.run.vm08.stdout:(127/136): python3-webob-1.8.8-2.el9.noarch.rpm 47 MB/s | 230 kB 00:00 2026-03-09T20:34:37.200 INFO:teuthology.orchestra.run.vm08.stdout:(128/136): python3-websocket-client-1.2.3-2.el9 26 MB/s | 90 kB 00:00 2026-03-09T20:34:37.202 INFO:teuthology.orchestra.run.vm08.stdout:(129/136): python3-xmltodict-0.12.0-15.el9.noar 8.4 MB/s | 22 kB 00:00 2026-03-09T20:34:37.207 INFO:teuthology.orchestra.run.vm08.stdout:(130/136): python3-werkzeug-2.0.3-3.el9.1.noarc 44 MB/s | 427 kB 00:00 2026-03-09T20:34:37.209 INFO:teuthology.orchestra.run.vm08.stdout:(131/136): python3-zc-lockfile-2.0-10.el9.noarc 3.2 MB/s | 20 kB 00:00 2026-03-09T20:34:37.212 INFO:teuthology.orchestra.run.vm08.stdout:(132/136): re2-20211101-20.el9.x86_64.rpm 40 MB/s | 191 kB 00:00 2026-03-09T20:34:37.234 INFO:teuthology.orchestra.run.vm08.stdout:(133/136): thrift-0.15.0-4.el9.x86_64.rpm 65 MB/s | 1.6 MB 00:00 2026-03-09T20:34:37.254 INFO:teuthology.orchestra.run.vm03.stdout:(73/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 1.7 MB/s | 442 kB 00:00 2026-03-09T20:34:37.299 INFO:teuthology.orchestra.run.vm03.stdout:(74/136): python3-packaging-20.9-5.el9.noarch.r 1.7 MB/s | 77 kB 00:00 2026-03-09T20:34:37.792 INFO:teuthology.orchestra.run.vm03.stdout:(75/136): python3-protobuf-3.14.0-17.el9.noarch 543 kB/s | 267 kB 00:00 2026-03-09T20:34:37.948 INFO:teuthology.orchestra.run.vm03.stdout:(76/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 1.0 MB/s | 157 kB 00:00 2026-03-09T20:34:37.965 INFO:teuthology.orchestra.run.vm03.stdout:(77/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 6.2 MB/s | 6.1 MB 00:00 2026-03-09T20:34:38.038 INFO:teuthology.orchestra.run.vm03.stdout:(78/136): python3-pyasn1-modules-0.4.8-7.el9.no 3.0 MB/s | 277 kB 00:00 2026-03-09T20:34:38.044 INFO:teuthology.orchestra.run.vm03.stdout:(79/136): python3-requests-oauthlib-1.3.0-12.el 680 kB/s | 54 kB 00:00 2026-03-09T20:34:38.124 INFO:teuthology.orchestra.run.vm03.stdout:(80/136): python3-toml-0.10.2-6.el9.noarch.rpm 523 kB/s | 42 kB 00:00 2026-03-09T20:34:38.239 INFO:teuthology.orchestra.run.vm03.stdout:(81/136): qatlib-25.08.0-2.el9.x86_64.rpm 2.0 MB/s | 240 kB 00:00 2026-03-09T20:34:38.298 INFO:teuthology.orchestra.run.vm03.stdout:(82/136): qatlib-service-25.08.0-2.el9.x86_64.r 623 kB/s | 37 kB 00:00 2026-03-09T20:34:38.353 INFO:teuthology.orchestra.run.vm03.stdout:(83/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 1.2 MB/s | 66 kB 00:00 2026-03-09T20:34:38.438 INFO:teuthology.orchestra.run.vm03.stdout:(84/136): socat-1.7.4.1-8.el9.x86_64.rpm 3.5 MB/s | 303 kB 00:00 2026-03-09T20:34:38.486 INFO:teuthology.orchestra.run.vm03.stdout:(85/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 1.3 MB/s | 64 kB 00:00 2026-03-09T20:34:38.604 INFO:teuthology.orchestra.run.vm08.stdout:(134/136): librados2-19.2.3-678.ge911bdeb.el9.x 2.5 MB/s | 3.4 MB 00:01 2026-03-09T20:34:38.809 INFO:teuthology.orchestra.run.vm03.stdout:(86/136): lua-devel-5.4.4-4.el9.x86_64.rpm 69 kB/s | 22 kB 00:00 2026-03-09T20:34:39.087 INFO:teuthology.orchestra.run.vm03.stdout:(87/136): protobuf-compiler-3.14.0-17.el9.x86_6 3.0 MB/s | 862 kB 00:00 2026-03-09T20:34:39.153 INFO:teuthology.orchestra.run.vm03.stdout:(88/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 8.2 MB/s | 551 kB 00:00 2026-03-09T20:34:39.156 INFO:teuthology.orchestra.run.vm04.stdout:(136/136): ceph-test-19.2.3-678.ge911bdeb.el9.x 3.5 MB/s | 50 MB 00:14 2026-03-09T20:34:39.159 INFO:teuthology.orchestra.run.vm04.stdout:-------------------------------------------------------------------------------- 2026-03-09T20:34:39.159 INFO:teuthology.orchestra.run.vm04.stdout:Total 9.5 MB/s | 210 MB 00:22 2026-03-09T20:34:39.169 INFO:teuthology.orchestra.run.vm03.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 19 MB/s | 308 kB 00:00 2026-03-09T20:34:39.172 INFO:teuthology.orchestra.run.vm03.stdout:(90/136): grpc-data-1.46.7-10.el9.noarch.rpm 9.0 MB/s | 19 kB 00:00 2026-03-09T20:34:39.182 INFO:teuthology.orchestra.run.vm08.stdout:(135/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 1.6 MB/s | 3.2 MB 00:01 2026-03-09T20:34:39.230 INFO:teuthology.orchestra.run.vm03.stdout:(91/136): libarrow-9.0.0-15.el9.x86_64.rpm 76 MB/s | 4.4 MB 00:00 2026-03-09T20:34:39.232 INFO:teuthology.orchestra.run.vm03.stdout:(92/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 11 MB/s | 25 kB 00:00 2026-03-09T20:34:39.235 INFO:teuthology.orchestra.run.vm03.stdout:(93/136): liboath-2.6.12-1.el9.x86_64.rpm 19 MB/s | 49 kB 00:00 2026-03-09T20:34:39.246 INFO:teuthology.orchestra.run.vm03.stdout:(94/136): libunwind-1.6.2-1.el9.x86_64.rpm 6.1 MB/s | 67 kB 00:00 2026-03-09T20:34:39.250 INFO:teuthology.orchestra.run.vm03.stdout:(95/136): luarocks-3.9.2-5.el9.noarch.rpm 36 MB/s | 151 kB 00:00 2026-03-09T20:34:39.262 INFO:teuthology.orchestra.run.vm03.stdout:(96/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 75 MB/s | 838 kB 00:00 2026-03-09T20:34:39.270 INFO:teuthology.orchestra.run.vm03.stdout:(97/136): python3-asyncssh-2.13.2-5.el9.noarch. 62 MB/s | 548 kB 00:00 2026-03-09T20:34:39.273 INFO:teuthology.orchestra.run.vm03.stdout:(98/136): python3-autocommand-2.2.2-8.el9.noarc 13 MB/s | 29 kB 00:00 2026-03-09T20:34:39.276 INFO:teuthology.orchestra.run.vm03.stdout:(99/136): python3-backports-tarfile-1.2.0-1.el9 23 MB/s | 60 kB 00:00 2026-03-09T20:34:39.279 INFO:teuthology.orchestra.run.vm03.stdout:(100/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 16 MB/s | 43 kB 00:00 2026-03-09T20:34:39.281 INFO:teuthology.orchestra.run.vm03.stdout:(101/136): python3-cachetools-4.2.4-1.el9.noarc 13 MB/s | 32 kB 00:00 2026-03-09T20:34:39.284 INFO:teuthology.orchestra.run.vm03.stdout:(102/136): python3-certifi-2023.05.07-4.el9.noa 6.6 MB/s | 14 kB 00:00 2026-03-09T20:34:39.288 INFO:teuthology.orchestra.run.vm03.stdout:(103/136): python3-cheroot-10.0.1-4.el9.noarch. 37 MB/s | 173 kB 00:00 2026-03-09T20:34:39.295 INFO:teuthology.orchestra.run.vm03.stdout:(104/136): python3-cherrypy-18.6.1-2.el9.noarch 54 MB/s | 358 kB 00:00 2026-03-09T20:34:39.301 INFO:teuthology.orchestra.run.vm03.stdout:(105/136): python3-google-auth-2.45.0-1.el9.noa 48 MB/s | 254 kB 00:00 2026-03-09T20:34:39.326 INFO:teuthology.orchestra.run.vm03.stdout:(106/136): python3-grpcio-1.46.7-10.el9.x86_64. 80 MB/s | 2.0 MB 00:00 2026-03-09T20:34:39.331 INFO:teuthology.orchestra.run.vm03.stdout:(107/136): python3-grpcio-tools-1.46.7-10.el9.x 35 MB/s | 144 kB 00:00 2026-03-09T20:34:39.333 INFO:teuthology.orchestra.run.vm03.stdout:(108/136): python3-jaraco-8.2.1-3.el9.noarch.rp 5.3 MB/s | 11 kB 00:00 2026-03-09T20:34:39.335 INFO:teuthology.orchestra.run.vm03.stdout:(109/136): python3-jaraco-classes-3.2.1-5.el9.n 8.8 MB/s | 18 kB 00:00 2026-03-09T20:34:39.337 INFO:teuthology.orchestra.run.vm03.stdout:(110/136): python3-jaraco-collections-3.0.0-8.e 11 MB/s | 23 kB 00:00 2026-03-09T20:34:39.349 INFO:teuthology.orchestra.run.vm03.stdout:(111/136): python3-jaraco-context-6.0.1-3.el9.n 1.6 MB/s | 20 kB 00:00 2026-03-09T20:34:39.351 INFO:teuthology.orchestra.run.vm03.stdout:(112/136): python3-jaraco-functools-3.5.0-2.el9 9.6 MB/s | 19 kB 00:00 2026-03-09T20:34:39.354 INFO:teuthology.orchestra.run.vm03.stdout:(113/136): python3-jaraco-text-4.0.0-2.el9.noar 9.7 MB/s | 26 kB 00:00 2026-03-09T20:34:39.367 INFO:teuthology.orchestra.run.vm03.stdout:(114/136): python3-kubernetes-26.1.0-3.el9.noar 79 MB/s | 1.0 MB 00:00 2026-03-09T20:34:39.370 INFO:teuthology.orchestra.run.vm03.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 14 MB/s | 46 kB 00:00 2026-03-09T20:34:39.373 INFO:teuthology.orchestra.run.vm03.stdout:(116/136): python3-more-itertools-8.12.0-2.el9. 28 MB/s | 79 kB 00:00 2026-03-09T20:34:39.376 INFO:teuthology.orchestra.run.vm03.stdout:(117/136): python3-natsort-7.1.1-5.el9.noarch.r 19 MB/s | 58 kB 00:00 2026-03-09T20:34:39.382 INFO:teuthology.orchestra.run.vm03.stdout:(118/136): python3-pecan-1.4.2-3.el9.noarch.rpm 53 MB/s | 272 kB 00:00 2026-03-09T20:34:39.386 INFO:teuthology.orchestra.run.vm03.stdout:(119/136): python3-portend-3.1.0-2.el9.noarch.r 4.1 MB/s | 16 kB 00:00 2026-03-09T20:34:39.389 INFO:teuthology.orchestra.run.vm03.stdout:(120/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 24 MB/s | 90 kB 00:00 2026-03-09T20:34:39.393 INFO:teuthology.orchestra.run.vm03.stdout:(121/136): python3-repoze-lru-0.7-16.el9.noarch 9.1 MB/s | 31 kB 00:00 2026-03-09T20:34:39.397 INFO:teuthology.orchestra.run.vm03.stdout:(122/136): python3-routes-2.5.1-5.el9.noarch.rp 45 MB/s | 188 kB 00:00 2026-03-09T20:34:39.400 INFO:teuthology.orchestra.run.vm03.stdout:(123/136): python3-rsa-4.9-2.el9.noarch.rpm 22 MB/s | 59 kB 00:00 2026-03-09T20:34:39.402 INFO:teuthology.orchestra.run.vm03.stdout:(124/136): python3-tempora-5.0.0-2.el9.noarch.r 15 MB/s | 36 kB 00:00 2026-03-09T20:34:39.406 INFO:teuthology.orchestra.run.vm03.stdout:(125/136): python3-typing-extensions-4.15.0-1.e 25 MB/s | 86 kB 00:00 2026-03-09T20:34:39.411 INFO:teuthology.orchestra.run.vm03.stdout:(126/136): python3-webob-1.8.8-2.el9.noarch.rpm 48 MB/s | 230 kB 00:00 2026-03-09T20:34:39.420 INFO:teuthology.orchestra.run.vm03.stdout:(127/136): python3-websocket-client-1.2.3-2.el9 10 MB/s | 90 kB 00:00 2026-03-09T20:34:39.427 INFO:teuthology.orchestra.run.vm03.stdout:(128/136): python3-werkzeug-2.0.3-3.el9.1.noarc 59 MB/s | 427 kB 00:00 2026-03-09T20:34:39.430 INFO:teuthology.orchestra.run.vm03.stdout:(129/136): python3-xmltodict-0.12.0-15.el9.noar 9.4 MB/s | 22 kB 00:00 2026-03-09T20:34:39.432 INFO:teuthology.orchestra.run.vm03.stdout:(130/136): python3-zc-lockfile-2.0-10.el9.noarc 9.5 MB/s | 20 kB 00:00 2026-03-09T20:34:39.436 INFO:teuthology.orchestra.run.vm03.stdout:(131/136): re2-20211101-20.el9.x86_64.rpm 46 MB/s | 191 kB 00:00 2026-03-09T20:34:39.458 INFO:teuthology.orchestra.run.vm03.stdout:(132/136): thrift-0.15.0-4.el9.x86_64.rpm 74 MB/s | 1.6 MB 00:00 2026-03-09T20:34:39.802 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T20:34:39.857 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T20:34:39.857 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T20:34:40.187 INFO:teuthology.orchestra.run.vm03.stdout:(133/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 9.0 MB/s | 19 MB 00:02 2026-03-09T20:34:40.784 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T20:34:40.784 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T20:34:40.896 INFO:teuthology.orchestra.run.vm03.stdout:(134/136): librados2-19.2.3-678.ge911bdeb.el9.x 2.4 MB/s | 3.4 MB 00:01 2026-03-09T20:34:41.729 INFO:teuthology.orchestra.run.vm03.stdout:(135/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 2.1 MB/s | 3.2 MB 00:01 2026-03-09T20:34:41.773 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T20:34:41.787 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-09T20:34:41.803 INFO:teuthology.orchestra.run.vm04.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-09T20:34:41.949 INFO:teuthology.orchestra.run.vm03.stdout:(136/136): ceph-test-19.2.3-678.ge911bdeb.el9.x 3.2 MB/s | 50 MB 00:15 2026-03-09T20:34:41.953 INFO:teuthology.orchestra.run.vm03.stdout:-------------------------------------------------------------------------------- 2026-03-09T20:34:41.953 INFO:teuthology.orchestra.run.vm03.stdout:Total 8.9 MB/s | 210 MB 00:23 2026-03-09T20:34:41.999 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-09T20:34:42.001 INFO:teuthology.orchestra.run.vm04.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:34:42.065 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:34:42.067 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T20:34:42.096 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T20:34:42.106 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T20:34:42.109 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-09T20:34:42.111 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-09T20:34:42.117 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-09T20:34:42.127 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-09T20:34:42.129 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:34:42.167 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:34:42.168 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T20:34:42.183 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T20:34:42.220 INFO:teuthology.orchestra.run.vm04.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-09T20:34:42.262 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-09T20:34:42.267 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-09T20:34:42.294 INFO:teuthology.orchestra.run.vm04.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-09T20:34:42.311 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-09T20:34:42.319 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-09T20:34:42.330 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-09T20:34:42.337 INFO:teuthology.orchestra.run.vm04.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-09T20:34:42.342 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-09T20:34:42.353 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-09T20:34:42.383 INFO:teuthology.orchestra.run.vm04.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-09T20:34:42.401 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-09T20:34:42.406 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-09T20:34:42.413 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-09T20:34:42.416 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-09T20:34:42.449 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-09T20:34:42.458 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-09T20:34:42.469 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-09T20:34:42.485 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-09T20:34:42.495 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-09T20:34:42.528 INFO:teuthology.orchestra.run.vm04.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-09T20:34:42.534 INFO:teuthology.orchestra.run.vm04.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-09T20:34:42.545 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-09T20:34:42.578 INFO:teuthology.orchestra.run.vm04.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-09T20:34:42.611 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T20:34:42.651 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-09T20:34:42.661 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T20:34:42.661 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T20:34:42.670 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-09T20:34:42.679 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-09T20:34:42.690 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-09T20:34:42.701 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-09T20:34:42.710 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-09T20:34:42.729 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-09T20:34:42.756 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-09T20:34:42.763 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-09T20:34:42.770 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-09T20:34:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-09T20:34:42.796 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-09T20:34:42.808 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-09T20:34:42.878 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-09T20:34:42.886 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-09T20:34:42.899 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-09T20:34:42.953 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-09T20:34:43.432 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-09T20:34:43.454 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-09T20:34:43.461 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-09T20:34:43.471 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-09T20:34:43.477 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-09T20:34:43.489 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-09T20:34:43.494 INFO:teuthology.orchestra.run.vm04.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-09T20:34:43.499 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-09T20:34:43.539 INFO:teuthology.orchestra.run.vm04.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-09T20:34:43.550 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T20:34:43.550 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T20:34:43.600 INFO:teuthology.orchestra.run.vm04.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-09T20:34:43.624 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-09T20:34:43.633 INFO:teuthology.orchestra.run.vm04.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-09T20:34:43.639 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-09T20:34:43.648 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-09T20:34:43.654 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-09T20:34:43.664 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-09T20:34:43.671 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-09T20:34:43.707 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-09T20:34:43.721 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-09T20:34:43.768 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-09T20:34:44.043 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-09T20:34:44.076 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-09T20:34:44.084 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-09T20:34:44.148 INFO:teuthology.orchestra.run.vm04.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-09T20:34:44.151 INFO:teuthology.orchestra.run.vm04.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-09T20:34:44.175 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-09T20:34:44.584 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-09T20:34:44.585 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T20:34:44.603 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-09T20:34:44.618 INFO:teuthology.orchestra.run.vm03.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-09T20:34:44.679 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-09T20:34:44.809 INFO:teuthology.orchestra.run.vm03.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-09T20:34:44.812 INFO:teuthology.orchestra.run.vm03.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:34:44.878 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:34:44.879 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T20:34:44.913 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T20:34:44.923 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T20:34:44.928 INFO:teuthology.orchestra.run.vm03.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-09T20:34:44.932 INFO:teuthology.orchestra.run.vm03.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-09T20:34:44.938 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-09T20:34:44.949 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-09T20:34:44.951 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:34:44.991 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:34:44.993 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T20:34:45.010 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T20:34:45.049 INFO:teuthology.orchestra.run.vm03.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-09T20:34:45.091 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-09T20:34:45.098 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-09T20:34:45.126 INFO:teuthology.orchestra.run.vm03.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-09T20:34:45.142 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-09T20:34:45.152 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-09T20:34:45.164 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-09T20:34:45.173 INFO:teuthology.orchestra.run.vm03.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-09T20:34:45.178 INFO:teuthology.orchestra.run.vm03.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-09T20:34:45.185 INFO:teuthology.orchestra.run.vm03.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-09T20:34:45.215 INFO:teuthology.orchestra.run.vm03.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-09T20:34:45.235 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-09T20:34:45.240 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-09T20:34:45.249 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-09T20:34:45.252 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-09T20:34:45.288 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-09T20:34:45.296 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-09T20:34:45.309 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-09T20:34:45.325 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-09T20:34:45.336 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-09T20:34:45.366 INFO:teuthology.orchestra.run.vm03.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-09T20:34:45.373 INFO:teuthology.orchestra.run.vm03.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-09T20:34:45.382 INFO:teuthology.orchestra.run.vm03.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-09T20:34:45.414 INFO:teuthology.orchestra.run.vm03.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-09T20:34:45.480 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-09T20:34:45.500 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-09T20:34:45.508 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-09T20:34:45.519 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-09T20:34:45.529 INFO:teuthology.orchestra.run.vm03.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-09T20:34:45.535 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-09T20:34:45.553 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-09T20:34:45.561 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-09T20:34:45.581 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-09T20:34:45.589 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-09T20:34:45.595 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-09T20:34:45.602 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-09T20:34:45.605 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-09T20:34:45.618 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-09T20:34:45.635 INFO:teuthology.orchestra.run.vm04.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-09T20:34:45.643 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-09T20:34:45.656 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-09T20:34:45.732 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-09T20:34:45.744 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-09T20:34:45.756 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-09T20:34:45.810 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-09T20:34:45.813 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-09T20:34:45.816 INFO:teuthology.orchestra.run.vm04.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T20:34:45.854 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T20:34:45.858 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-09T20:34:45.867 INFO:teuthology.orchestra.run.vm04.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-09T20:34:46.146 INFO:teuthology.orchestra.run.vm04.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-09T20:34:46.149 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T20:34:46.175 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T20:34:46.178 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-09T20:34:46.247 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-09T20:34:46.265 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-09T20:34:46.272 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-09T20:34:46.281 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-09T20:34:46.287 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-09T20:34:46.295 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-09T20:34:46.299 INFO:teuthology.orchestra.run.vm03.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-09T20:34:46.302 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-09T20:34:46.334 INFO:teuthology.orchestra.run.vm03.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-09T20:34:46.392 INFO:teuthology.orchestra.run.vm03.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-09T20:34:46.407 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-09T20:34:46.417 INFO:teuthology.orchestra.run.vm03.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-09T20:34:46.423 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-09T20:34:46.434 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-09T20:34:46.440 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-09T20:34:46.455 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-09T20:34:46.461 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-09T20:34:46.496 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-09T20:34:46.512 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-09T20:34:46.562 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-09T20:34:46.866 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-09T20:34:46.901 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-09T20:34:46.909 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-09T20:34:46.974 INFO:teuthology.orchestra.run.vm03.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-09T20:34:46.977 INFO:teuthology.orchestra.run.vm03.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-09T20:34:47.003 INFO:teuthology.orchestra.run.vm03.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-09T20:34:47.367 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:34:47.405 INFO:teuthology.orchestra.run.vm03.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-09T20:34:47.490 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:34:47.517 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:34:47.537 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-09T20:34:47.559 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-09T20:34:47.571 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-09T20:34:47.667 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-09T20:34:47.683 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-09T20:34:47.716 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-09T20:34:47.757 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-09T20:34:47.821 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-09T20:34:47.832 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-09T20:34:47.839 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T20:34:47.846 INFO:teuthology.orchestra.run.vm04.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-09T20:34:47.851 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-09T20:34:47.853 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T20:34:47.877 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T20:34:48.212 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-09T20:34:48.274 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T20:34:48.327 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T20:34:48.327 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-09T20:34:48.327 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T20:34:48.327 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:34:48.333 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T20:34:48.439 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-09T20:34:48.468 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-09T20:34:48.475 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-09T20:34:48.481 INFO:teuthology.orchestra.run.vm03.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-09T20:34:48.651 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-09T20:34:48.654 INFO:teuthology.orchestra.run.vm03.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T20:34:48.688 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T20:34:48.693 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-09T20:34:48.702 INFO:teuthology.orchestra.run.vm03.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-09T20:34:48.975 INFO:teuthology.orchestra.run.vm03.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-09T20:34:48.980 INFO:teuthology.orchestra.run.vm03.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T20:34:49.002 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T20:34:49.073 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-09T20:34:50.309 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:34:50.498 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:34:50.527 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:34:50.605 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-09T20:34:50.628 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-09T20:34:50.736 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-09T20:34:50.761 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-09T20:34:50.796 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-09T20:34:50.847 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-09T20:34:50.920 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-09T20:34:50.937 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-09T20:34:50.943 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T20:34:50.950 INFO:teuthology.orchestra.run.vm03.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-09T20:34:50.954 INFO:teuthology.orchestra.run.vm03.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-09T20:34:50.957 INFO:teuthology.orchestra.run.vm03.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T20:34:50.983 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T20:34:51.337 INFO:teuthology.orchestra.run.vm03.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-09T20:34:51.346 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T20:34:51.394 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T20:34:51.394 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-09T20:34:51.394 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T20:34:51.394 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:34:51.396 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T20:34:55.318 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T20:34:55.318 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /sys 2026-03-09T20:34:55.318 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /proc 2026-03-09T20:34:55.319 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /mnt 2026-03-09T20:34:55.319 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /var/tmp 2026-03-09T20:34:55.319 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /home 2026-03-09T20:34:55.319 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /root 2026-03-09T20:34:55.319 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /tmp 2026-03-09T20:34:55.319 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:34:55.453 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T20:34:55.484 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T20:34:55.484 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:34:55.484 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T20:34:55.484 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T20:34:55.484 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T20:34:55.484 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:34:55.742 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T20:34:55.766 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T20:34:55.766 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:34:55.766 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T20:34:55.766 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T20:34:55.766 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T20:34:55.766 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:34:55.782 INFO:teuthology.orchestra.run.vm04.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-09T20:34:55.793 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-09T20:34:55.814 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:34:55.814 INFO:teuthology.orchestra.run.vm04.stdout:Creating group 'qat' with GID 994. 2026-03-09T20:34:55.814 INFO:teuthology.orchestra.run.vm04.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T20:34:55.814 INFO:teuthology.orchestra.run.vm04.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T20:34:55.814 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:34:55.827 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:34:55.857 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:34:55.857 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T20:34:55.857 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:34:55.901 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-09T20:34:55.979 INFO:teuthology.orchestra.run.vm04.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-09T20:34:55.985 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T20:34:56.004 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T20:34:56.004 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:34:56.004 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T20:34:56.004 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:34:56.849 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T20:34:56.879 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T20:34:56.879 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:34:56.879 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T20:34:56.879 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T20:34:56.879 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T20:34:56.879 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:34:56.948 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T20:34:56.953 INFO:teuthology.orchestra.run.vm04.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T20:34:56.960 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-09T20:34:56.984 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-09T20:34:56.987 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T20:34:57.538 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T20:34:57.737 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T20:34:58.295 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T20:34:58.301 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T20:34:58.313 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T20:34:58.313 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /sys 2026-03-09T20:34:58.313 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /proc 2026-03-09T20:34:58.313 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /mnt 2026-03-09T20:34:58.313 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /var/tmp 2026-03-09T20:34:58.313 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /home 2026-03-09T20:34:58.313 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /root 2026-03-09T20:34:58.313 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /tmp 2026-03-09T20:34:58.313 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:34:58.374 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T20:34:58.464 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-09T20:34:58.488 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T20:34:58.490 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T20:34:58.516 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T20:34:58.516 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:34:58.516 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T20:34:58.516 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T20:34:58.516 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T20:34:58.516 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:34:58.516 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T20:34:58.516 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:34:58.516 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T20:34:58.516 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T20:34:58.516 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T20:34:58.516 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:34:58.667 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T20:34:58.679 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T20:34:58.761 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T20:34:58.786 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T20:34:58.786 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:34:58.786 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T20:34:58.786 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T20:34:58.787 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T20:34:58.787 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:34:58.796 INFO:teuthology.orchestra.run.vm03.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-09T20:34:58.799 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-09T20:34:58.819 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:34:58.819 INFO:teuthology.orchestra.run.vm03.stdout:Creating group 'qat' with GID 994. 2026-03-09T20:34:58.819 INFO:teuthology.orchestra.run.vm03.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T20:34:58.819 INFO:teuthology.orchestra.run.vm03.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T20:34:58.819 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:34:58.833 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:34:58.864 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:34:58.864 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T20:34:58.864 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:34:58.909 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-09T20:34:58.988 INFO:teuthology.orchestra.run.vm03.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-09T20:34:58.993 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T20:34:59.009 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T20:34:59.009 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:34:59.009 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T20:34:59.009 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:34:59.217 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-09T20:34:59.220 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T20:34:59.245 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T20:34:59.245 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:34:59.245 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T20:34:59.245 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T20:34:59.245 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T20:34:59.245 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:34:59.257 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T20:34:59.281 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T20:34:59.281 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:34:59.281 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T20:34:59.281 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:34:59.439 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T20:34:59.465 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T20:34:59.465 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:34:59.465 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T20:34:59.465 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T20:34:59.465 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T20:34:59.465 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:34:59.599 INFO:teuthology.orchestra.run.vm08.stdout:(136/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 840 kB/s | 19 MB 00:23 2026-03-09T20:34:59.604 INFO:teuthology.orchestra.run.vm08.stdout:-------------------------------------------------------------------------------- 2026-03-09T20:34:59.604 INFO:teuthology.orchestra.run.vm08.stdout:Total 5.4 MB/s | 210 MB 00:38 2026-03-09T20:34:59.871 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T20:34:59.899 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T20:34:59.900 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:34:59.900 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T20:34:59.900 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T20:34:59.900 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T20:34:59.900 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:34:59.968 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T20:34:59.972 INFO:teuthology.orchestra.run.vm03.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T20:34:59.978 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-09T20:35:00.002 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-09T20:35:00.005 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T20:35:00.134 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:35:00.184 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:35:00.184 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:35:00.576 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T20:35:00.582 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T20:35:01.045 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:35:01.045 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:35:01.129 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T20:35:01.131 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T20:35:01.202 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T20:35:01.264 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-09T20:35:01.266 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T20:35:01.289 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T20:35:01.289 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:35:01.289 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T20:35:01.289 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T20:35:01.289 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T20:35:01.289 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:35:01.305 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T20:35:01.318 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T20:35:01.855 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-09T20:35:01.858 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T20:35:01.884 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T20:35:01.884 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:35:01.884 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T20:35:01.884 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T20:35:01.884 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T20:35:01.884 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:35:01.900 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T20:35:01.925 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T20:35:01.925 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:35:01.925 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T20:35:01.925 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:35:01.975 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:35:01.990 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-09T20:35:02.004 INFO:teuthology.orchestra.run.vm08.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-09T20:35:02.090 INFO:teuthology.orchestra.run.vm03.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T20:35:02.114 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T20:35:02.114 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:35:02.114 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T20:35:02.114 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T20:35:02.115 INFO:teuthology.orchestra.run.vm03.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T20:35:02.115 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:35:02.151 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-09T20:35:02.163 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-09T20:35:02.169 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-09T20:35:02.182 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-09T20:35:02.184 INFO:teuthology.orchestra.run.vm08.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:35:02.229 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-09T20:35:02.240 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T20:35:02.244 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-09T20:35:02.244 INFO:teuthology.orchestra.run.vm04.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T20:35:02.248 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:35:02.250 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T20:35:02.265 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T20:35:02.265 INFO:teuthology.orchestra.run.vm04.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:35:02.282 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T20:35:02.293 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T20:35:02.297 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-09T20:35:02.299 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-09T20:35:02.305 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-09T20:35:02.316 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-09T20:35:02.318 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:35:02.356 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:35:02.358 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T20:35:02.376 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T20:35:02.421 INFO:teuthology.orchestra.run.vm08.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-09T20:35:02.470 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-09T20:35:02.477 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-09T20:35:02.502 INFO:teuthology.orchestra.run.vm08.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-09T20:35:02.517 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-09T20:35:02.525 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-09T20:35:02.538 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-09T20:35:02.546 INFO:teuthology.orchestra.run.vm08.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-09T20:35:02.550 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-09T20:35:02.556 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-09T20:35:02.587 INFO:teuthology.orchestra.run.vm08.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-09T20:35:02.605 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-09T20:35:02.609 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-09T20:35:02.618 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-09T20:35:02.621 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-09T20:35:02.653 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-09T20:35:02.660 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-09T20:35:02.670 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-09T20:35:02.684 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-09T20:35:02.692 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-09T20:35:02.722 INFO:teuthology.orchestra.run.vm08.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-09T20:35:02.728 INFO:teuthology.orchestra.run.vm08.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-09T20:35:02.748 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-09T20:35:02.785 INFO:teuthology.orchestra.run.vm08.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-09T20:35:02.861 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-09T20:35:02.986 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-09T20:35:02.999 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-09T20:35:03.010 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-09T20:35:03.018 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-09T20:35:03.024 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-09T20:35:03.047 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-09T20:35:03.128 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-09T20:35:03.206 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-09T20:35:03.236 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-09T20:35:03.253 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-09T20:35:03.267 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-09T20:35:03.281 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-09T20:35:03.416 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-09T20:35:03.448 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-09T20:35:03.480 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-09T20:35:03.535 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-09T20:35:03.742 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:35:03.742 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-09T20:35:03.742 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-09T20:35:03.742 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-09T20:35:03.742 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:35:03.742 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-09T20:35:03.742 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T20:35:03.742 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-09T20:35:03.742 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-09T20:35:03.742 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-09T20:35:03.742 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-09T20:35:03.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-09T20:35:03.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-09T20:35:03.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-09T20:35:03.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-09T20:35:03.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-09T20:35:03.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-09T20:35:03.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-09T20:35:03.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-09T20:35:03.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-09T20:35:03.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-09T20:35:03.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-09T20:35:03.750 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout:Upgraded: 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout:Installed: 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T20:35:03.868 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T20:35:03.869 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T20:35:03.870 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T20:35:03.871 INFO:teuthology.orchestra.run.vm04.stdout: zip-3.0-35.el9.x86_64 2026-03-09T20:35:03.871 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:35:03.871 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:35:03.974 DEBUG:teuthology.parallel:result is None 2026-03-09T20:35:03.980 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-09T20:35:03.998 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-09T20:35:04.005 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-09T20:35:04.014 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-09T20:35:04.020 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-09T20:35:04.030 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-09T20:35:04.035 INFO:teuthology.orchestra.run.vm08.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-09T20:35:04.038 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-09T20:35:04.070 INFO:teuthology.orchestra.run.vm08.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-09T20:35:04.128 INFO:teuthology.orchestra.run.vm08.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-09T20:35:04.144 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-09T20:35:04.154 INFO:teuthology.orchestra.run.vm08.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-09T20:35:04.160 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-09T20:35:04.170 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-09T20:35:04.177 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-09T20:35:04.188 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-09T20:35:04.195 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-09T20:35:04.231 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-09T20:35:04.321 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-09T20:35:04.369 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-09T20:35:04.698 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-09T20:35:04.882 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-09T20:35:05.026 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-09T20:35:05.028 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-09T20:35:05.042 INFO:teuthology.orchestra.run.vm03.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-09T20:35:05.048 INFO:teuthology.orchestra.run.vm03.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-09T20:35:05.103 INFO:teuthology.orchestra.run.vm08.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-09T20:35:05.108 INFO:teuthology.orchestra.run.vm08.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-09T20:35:05.112 INFO:teuthology.orchestra.run.vm03.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-09T20:35:05.123 INFO:teuthology.orchestra.run.vm03.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T20:35:05.127 INFO:teuthology.orchestra.run.vm03.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-09T20:35:05.127 INFO:teuthology.orchestra.run.vm03.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T20:35:05.138 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-09T20:35:05.145 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T20:35:05.145 INFO:teuthology.orchestra.run.vm03.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:35:05.544 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-09T20:35:05.659 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-09T20:35:06.514 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-09T20:35:06.515 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-09T20:35:06.516 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-09T20:35:06.517 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-09T20:35:06.518 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-09T20:35:06.518 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-09T20:35:06.518 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-09T20:35:06.518 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-09T20:35:06.518 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-09T20:35:06.518 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-09T20:35:06.518 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-09T20:35:06.518 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-09T20:35:06.518 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-09T20:35:06.518 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-09T20:35:06.518 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-09T20:35:06.518 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-09T20:35:06.519 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-09T20:35:06.520 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-09T20:35:06.520 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-09T20:35:06.520 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-09T20:35:06.520 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-09T20:35:06.520 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-09T20:35:06.520 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-09T20:35:06.520 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-09T20:35:06.520 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-09T20:35:06.520 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-09T20:35:06.520 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T20:35:06.520 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-09T20:35:06.520 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-09T20:35:06.618 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout:Upgraded: 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout:Installed: 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T20:35:06.632 INFO:teuthology.orchestra.run.vm03.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T20:35:06.633 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T20:35:06.634 INFO:teuthology.orchestra.run.vm03.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T20:35:06.635 INFO:teuthology.orchestra.run.vm03.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T20:35:06.635 INFO:teuthology.orchestra.run.vm03.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.635 INFO:teuthology.orchestra.run.vm03.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.635 INFO:teuthology.orchestra.run.vm03.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:06.635 INFO:teuthology.orchestra.run.vm03.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T20:35:06.635 INFO:teuthology.orchestra.run.vm03.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T20:35:06.635 INFO:teuthology.orchestra.run.vm03.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T20:35:06.635 INFO:teuthology.orchestra.run.vm03.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T20:35:06.635 INFO:teuthology.orchestra.run.vm03.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T20:35:06.635 INFO:teuthology.orchestra.run.vm03.stdout: zip-3.0-35.el9.x86_64 2026-03-09T20:35:06.635 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:35:06.635 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:35:06.652 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-09T20:35:06.660 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-09T20:35:06.665 INFO:teuthology.orchestra.run.vm08.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-09T20:35:06.731 DEBUG:teuthology.parallel:result is None 2026-03-09T20:35:06.832 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-09T20:35:06.835 INFO:teuthology.orchestra.run.vm08.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T20:35:06.869 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T20:35:06.874 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-09T20:35:06.883 INFO:teuthology.orchestra.run.vm08.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-09T20:35:07.154 INFO:teuthology.orchestra.run.vm08.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-09T20:35:07.157 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T20:35:07.179 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T20:35:07.182 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-09T20:35:08.333 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:35:08.341 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:35:08.365 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T20:35:08.384 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-09T20:35:08.408 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-09T20:35:08.509 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-09T20:35:08.524 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-09T20:35:08.555 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-09T20:35:08.616 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-09T20:35:08.679 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-09T20:35:08.689 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-09T20:35:08.695 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T20:35:08.700 INFO:teuthology.orchestra.run.vm08.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-09T20:35:08.705 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-09T20:35:08.707 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T20:35:08.727 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T20:35:09.045 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-09T20:35:09.054 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T20:35:09.107 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T20:35:09.107 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-09T20:35:09.107 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T20:35:09.107 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:35:09.112 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T20:35:15.938 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T20:35:15.938 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /sys 2026-03-09T20:35:15.938 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /proc 2026-03-09T20:35:15.938 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /mnt 2026-03-09T20:35:15.938 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /var/tmp 2026-03-09T20:35:15.938 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /home 2026-03-09T20:35:15.938 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /root 2026-03-09T20:35:15.938 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /tmp 2026-03-09T20:35:15.938 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:35:16.067 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T20:35:16.093 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T20:35:16.094 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:35:16.094 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T20:35:16.094 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T20:35:16.094 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T20:35:16.094 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:35:16.325 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T20:35:16.350 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T20:35:16.350 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:35:16.350 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T20:35:16.350 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T20:35:16.350 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T20:35:16.350 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:35:16.360 INFO:teuthology.orchestra.run.vm08.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-09T20:35:16.363 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-09T20:35:16.382 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:35:16.382 INFO:teuthology.orchestra.run.vm08.stdout:Creating group 'qat' with GID 994. 2026-03-09T20:35:16.382 INFO:teuthology.orchestra.run.vm08.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T20:35:16.382 INFO:teuthology.orchestra.run.vm08.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T20:35:16.382 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:35:16.393 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:35:16.422 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T20:35:16.422 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T20:35:16.422 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:35:16.464 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-09T20:35:16.540 INFO:teuthology.orchestra.run.vm08.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-09T20:35:16.545 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T20:35:16.560 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T20:35:16.560 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:35:16.560 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T20:35:16.560 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:35:17.388 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T20:35:17.416 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T20:35:17.416 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:35:17.416 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T20:35:17.416 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T20:35:17.416 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T20:35:17.416 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:35:17.481 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T20:35:17.484 INFO:teuthology.orchestra.run.vm08.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T20:35:17.491 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-09T20:35:17.515 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-09T20:35:17.519 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T20:35:18.081 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T20:35:18.088 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T20:35:18.648 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T20:35:18.651 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T20:35:18.719 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T20:35:18.778 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-09T20:35:18.781 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T20:35:18.805 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T20:35:18.805 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:35:18.805 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T20:35:18.805 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T20:35:18.805 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T20:35:18.805 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:35:18.821 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T20:35:18.835 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T20:35:19.361 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-09T20:35:19.416 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T20:35:19.441 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T20:35:19.441 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:35:19.441 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T20:35:19.441 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T20:35:19.441 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T20:35:19.441 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:35:19.466 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T20:35:19.485 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T20:35:19.485 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:35:19.485 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T20:35:19.485 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:35:19.661 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T20:35:19.686 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T20:35:19.686 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:35:19.686 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T20:35:19.686 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T20:35:19.686 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T20:35:19.686 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:35:22.304 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-09T20:35:22.315 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-09T20:35:22.321 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-09T20:35:22.376 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-09T20:35:22.386 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T20:35:22.390 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-09T20:35:22.390 INFO:teuthology.orchestra.run.vm08.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T20:35:22.408 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T20:35:22.408 INFO:teuthology.orchestra.run.vm08.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:35:23.782 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:35:23.782 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-09T20:35:23.782 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-09T20:35:23.782 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-09T20:35:23.782 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T20:35:23.782 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-09T20:35:23.782 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T20:35:23.782 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-09T20:35:23.782 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-09T20:35:23.782 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-09T20:35:23.782 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-09T20:35:23.782 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T20:35:23.783 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-09T20:35:23.783 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-09T20:35:23.783 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-09T20:35:23.783 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-09T20:35:23.783 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-09T20:35:23.783 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-09T20:35:23.783 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-09T20:35:23.783 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-09T20:35:23.783 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-09T20:35:23.783 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-09T20:35:23.783 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-09T20:35:23.784 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-09T20:35:23.784 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-09T20:35:23.784 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-09T20:35:23.784 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-09T20:35:23.784 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-09T20:35:23.784 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-09T20:35:23.784 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-09T20:35:23.784 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-09T20:35:23.784 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-09T20:35:23.784 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-09T20:35:23.784 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-09T20:35:23.784 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-09T20:35:23.784 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-09T20:35:23.785 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-09T20:35:23.786 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-09T20:35:23.787 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout:Upgraded: 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout:Installed: 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:23.895 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T20:35:23.896 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T20:35:23.897 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout: zip-3.0-35.el9.x86_64 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:35:23.898 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:35:23.989 DEBUG:teuthology.parallel:result is None 2026-03-09T20:35:23.990 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:35:24.622 DEBUG:teuthology.orchestra.run.vm03:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T20:35:24.645 INFO:teuthology.orchestra.run.vm03.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T20:35:24.646 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T20:35:24.646 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T20:35:24.647 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:35:25.221 DEBUG:teuthology.orchestra.run.vm04:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T20:35:25.242 INFO:teuthology.orchestra.run.vm04.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T20:35:25.243 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T20:35:25.243 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T20:35:25.243 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:35:25.818 DEBUG:teuthology.orchestra.run.vm08:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T20:35:25.839 INFO:teuthology.orchestra.run.vm08.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T20:35:25.840 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T20:35:25.840 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T20:35:25.841 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-09T20:35:25.841 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T20:35:25.841 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T20:35:25.872 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T20:35:25.872 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T20:35:25.903 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:35:25.903 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T20:35:25.931 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-09T20:35:25.932 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T20:35:25.932 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T20:35:25.958 DEBUG:teuthology.orchestra.run.vm03:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T20:35:26.025 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T20:35:26.025 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T20:35:26.051 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T20:35:26.115 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:35:26.115 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T20:35:26.139 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T20:35:26.204 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-09T20:35:26.204 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T20:35:26.204 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T20:35:26.229 DEBUG:teuthology.orchestra.run.vm03:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T20:35:26.292 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T20:35:26.292 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T20:35:26.318 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T20:35:26.386 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:35:26.386 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T20:35:26.413 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T20:35:26.481 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-09T20:35:26.481 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T20:35:26.481 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T20:35:26.507 DEBUG:teuthology.orchestra.run.vm03:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T20:35:26.575 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T20:35:26.575 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T20:35:26.603 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T20:35:26.672 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:35:26.672 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T20:35:26.697 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T20:35:26.763 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-09T20:35:26.806 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon election default strategy': 3}, 'mgr': {'debug mgr': 20, 'debug ms': 1, 'mgr/cephadm/use_agent': True}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'MON_DOWN', 'mons down', 'mon down', 'out of quorum', 'CEPHADM_STRAY_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-09T20:35:26.806 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:35:26.807 INFO:tasks.cephadm:Cluster fsid is 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:35:26.807 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-09T20:35:26.807 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.103', 'mon.b': '192.168.123.104', 'mon.c': '192.168.123.108'} 2026-03-09T20:35:26.807 INFO:tasks.cephadm:First mon is mon.a on vm03 2026-03-09T20:35:26.807 INFO:tasks.cephadm:First mgr is a 2026-03-09T20:35:26.807 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-09T20:35:26.807 DEBUG:teuthology.orchestra.run.vm03:> sudo hostname $(hostname -s) 2026-03-09T20:35:26.831 DEBUG:teuthology.orchestra.run.vm04:> sudo hostname $(hostname -s) 2026-03-09T20:35:26.858 DEBUG:teuthology.orchestra.run.vm08:> sudo hostname $(hostname -s) 2026-03-09T20:35:26.883 INFO:tasks.cephadm:Downloading "compiled" cephadm from cachra 2026-03-09T20:35:26.883 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:35:27.481 INFO:tasks.cephadm:builder_project result: [{'url': 'https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'chacra_url': 'https://3.chacra.ceph.com/repos/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'ref': 'squid', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'distro': 'centos', 'distro_version': '9', 'distro_codename': None, 'modified': '2026-02-25 18:55:15.146628', 'status': 'ready', 'flavor': 'default', 'project': 'ceph', 'archs': ['source', 'x86_64'], 'extra': {'version': '19.2.3-678-ge911bdeb', 'package_manager_version': '19.2.3-678.ge911bdeb', 'build_url': 'https://jenkins.ceph.com/job/ceph-dev-pipeline/3275/', 'root_build_cause': '', 'node_name': '10.20.192.26+soko16', 'job_name': 'ceph-dev-pipeline'}}] 2026-03-09T20:35:28.138 INFO:tasks.util.chacra:got chacra host 3.chacra.ceph.com, ref squid, sha1 e911bdebe5c8faa3800735d1568fcdca65db60df from https://shaman.ceph.com/api/search/?project=ceph&distros=centos%2F9%2Fx86_64&flavor=default&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:35:28.139 INFO:tasks.cephadm:Discovered cachra url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-09T20:35:28.139 INFO:tasks.cephadm:Downloading cephadm from url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-09T20:35:28.139 DEBUG:teuthology.orchestra.run.vm03:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T20:35:29.480 INFO:teuthology.orchestra.run.vm03.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 20:35 /home/ubuntu/cephtest/cephadm 2026-03-09T20:35:29.480 DEBUG:teuthology.orchestra.run.vm04:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T20:35:30.869 INFO:teuthology.orchestra.run.vm04.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 20:35 /home/ubuntu/cephtest/cephadm 2026-03-09T20:35:30.869 DEBUG:teuthology.orchestra.run.vm08:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T20:35:32.275 INFO:teuthology.orchestra.run.vm08.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 20:35 /home/ubuntu/cephtest/cephadm 2026-03-09T20:35:32.275 DEBUG:teuthology.orchestra.run.vm03:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T20:35:32.294 DEBUG:teuthology.orchestra.run.vm04:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T20:35:32.309 DEBUG:teuthology.orchestra.run.vm08:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T20:35:32.330 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df on all hosts... 2026-03-09T20:35:32.330 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T20:35:32.336 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T20:35:32.351 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T20:35:32.507 INFO:teuthology.orchestra.run.vm03.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T20:35:32.518 INFO:teuthology.orchestra.run.vm04.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T20:35:32.539 INFO:teuthology.orchestra.run.vm08.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T20:37:02.253 INFO:teuthology.orchestra.run.vm04.stdout:{ 2026-03-09T20:37:02.253 INFO:teuthology.orchestra.run.vm04.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T20:37:02.253 INFO:teuthology.orchestra.run.vm04.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T20:37:02.253 INFO:teuthology.orchestra.run.vm04.stdout: "repo_digests": [ 2026-03-09T20:37:02.253 INFO:teuthology.orchestra.run.vm04.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T20:37:02.253 INFO:teuthology.orchestra.run.vm04.stdout: ] 2026-03-09T20:37:02.253 INFO:teuthology.orchestra.run.vm04.stdout:} 2026-03-09T20:37:11.172 INFO:teuthology.orchestra.run.vm08.stdout:{ 2026-03-09T20:37:11.173 INFO:teuthology.orchestra.run.vm08.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T20:37:11.173 INFO:teuthology.orchestra.run.vm08.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T20:37:11.173 INFO:teuthology.orchestra.run.vm08.stdout: "repo_digests": [ 2026-03-09T20:37:11.173 INFO:teuthology.orchestra.run.vm08.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T20:37:11.173 INFO:teuthology.orchestra.run.vm08.stdout: ] 2026-03-09T20:37:11.173 INFO:teuthology.orchestra.run.vm08.stdout:} 2026-03-09T20:37:11.216 INFO:teuthology.orchestra.run.vm03.stdout:{ 2026-03-09T20:37:11.216 INFO:teuthology.orchestra.run.vm03.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T20:37:11.216 INFO:teuthology.orchestra.run.vm03.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T20:37:11.216 INFO:teuthology.orchestra.run.vm03.stdout: "repo_digests": [ 2026-03-09T20:37:11.216 INFO:teuthology.orchestra.run.vm03.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T20:37:11.216 INFO:teuthology.orchestra.run.vm03.stdout: ] 2026-03-09T20:37:11.216 INFO:teuthology.orchestra.run.vm03.stdout:} 2026-03-09T20:37:11.233 DEBUG:teuthology.orchestra.run.vm03:> sudo mkdir -p /etc/ceph 2026-03-09T20:37:11.259 DEBUG:teuthology.orchestra.run.vm04:> sudo mkdir -p /etc/ceph 2026-03-09T20:37:11.285 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /etc/ceph 2026-03-09T20:37:11.317 DEBUG:teuthology.orchestra.run.vm03:> sudo chmod 777 /etc/ceph 2026-03-09T20:37:11.347 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod 777 /etc/ceph 2026-03-09T20:37:11.370 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 777 /etc/ceph 2026-03-09T20:37:11.401 INFO:tasks.cephadm:Writing seed config... 2026-03-09T20:37:11.401 INFO:tasks.cephadm: override: [global] mon election default strategy = 3 2026-03-09T20:37:11.401 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-09T20:37:11.401 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-09T20:37:11.401 INFO:tasks.cephadm: override: [mgr] mgr/cephadm/use_agent = True 2026-03-09T20:37:11.401 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-09T20:37:11.401 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-09T20:37:11.401 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-09T20:37:11.402 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-09T20:37:11.402 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-09T20:37:11.402 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-09T20:37:11.402 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T20:37:11.402 DEBUG:teuthology.orchestra.run.vm03:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-09T20:37:11.421 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 8176208c-1bf7-11f1-86ba-4f626c793dbd mon election default strategy = 3 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 mgr/cephadm/use_agent = True [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-09T20:37:11.421 DEBUG:teuthology.orchestra.run.vm03:mon.a> sudo journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mon.a.service 2026-03-09T20:37:11.463 DEBUG:teuthology.orchestra.run.vm03:mgr.a> sudo journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mgr.a.service 2026-03-09T20:37:11.505 INFO:tasks.cephadm:Bootstrapping... 2026-03-09T20:37:11.505 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df -v bootstrap --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id a --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.103 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:37:11.663 INFO:teuthology.orchestra.run.vm03.stdout:-------------------------------------------------------------------------------- 2026-03-09T20:37:11.663 INFO:teuthology.orchestra.run.vm03.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df', '-v', 'bootstrap', '--fsid', '8176208c-1bf7-11f1-86ba-4f626c793dbd', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'a', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.103', '--skip-admin-label'] 2026-03-09T20:37:11.663 INFO:teuthology.orchestra.run.vm03.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-09T20:37:11.663 INFO:teuthology.orchestra.run.vm03.stdout:Verifying podman|docker is present... 2026-03-09T20:37:11.686 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stdout 5.8.0 2026-03-09T20:37:11.686 INFO:teuthology.orchestra.run.vm03.stdout:Verifying lvm2 is present... 2026-03-09T20:37:11.686 INFO:teuthology.orchestra.run.vm03.stdout:Verifying time synchronization is in place... 2026-03-09T20:37:11.694 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-09T20:37:11.695 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-09T20:37:11.702 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-09T20:37:11.702 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout inactive 2026-03-09T20:37:11.709 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout enabled 2026-03-09T20:37:11.716 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout active 2026-03-09T20:37:11.716 INFO:teuthology.orchestra.run.vm03.stdout:Unit chronyd.service is enabled and running 2026-03-09T20:37:11.716 INFO:teuthology.orchestra.run.vm03.stdout:Repeating the final host check... 2026-03-09T20:37:11.736 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stdout 5.8.0 2026-03-09T20:37:11.736 INFO:teuthology.orchestra.run.vm03.stdout:podman (/bin/podman) version 5.8.0 is present 2026-03-09T20:37:11.736 INFO:teuthology.orchestra.run.vm03.stdout:systemctl is present 2026-03-09T20:37:11.736 INFO:teuthology.orchestra.run.vm03.stdout:lvcreate is present 2026-03-09T20:37:11.742 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-09T20:37:11.742 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-09T20:37:11.749 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-09T20:37:11.749 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout inactive 2026-03-09T20:37:11.757 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout enabled 2026-03-09T20:37:11.763 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stdout active 2026-03-09T20:37:11.763 INFO:teuthology.orchestra.run.vm03.stdout:Unit chronyd.service is enabled and running 2026-03-09T20:37:11.763 INFO:teuthology.orchestra.run.vm03.stdout:Host looks OK 2026-03-09T20:37:11.763 INFO:teuthology.orchestra.run.vm03.stdout:Cluster fsid: 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:11.764 INFO:teuthology.orchestra.run.vm03.stdout:Acquiring lock 140438950426512 on /run/cephadm/8176208c-1bf7-11f1-86ba-4f626c793dbd.lock 2026-03-09T20:37:11.764 INFO:teuthology.orchestra.run.vm03.stdout:Lock 140438950426512 acquired on /run/cephadm/8176208c-1bf7-11f1-86ba-4f626c793dbd.lock 2026-03-09T20:37:11.764 INFO:teuthology.orchestra.run.vm03.stdout:Verifying IP 192.168.123.103 port 3300 ... 2026-03-09T20:37:11.764 INFO:teuthology.orchestra.run.vm03.stdout:Verifying IP 192.168.123.103 port 6789 ... 2026-03-09T20:37:11.764 INFO:teuthology.orchestra.run.vm03.stdout:Base mon IP(s) is [192.168.123.103:3300, 192.168.123.103:6789], mon addrv is [v2:192.168.123.103:3300,v1:192.168.123.103:6789] 2026-03-09T20:37:11.769 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.103 metric 100 2026-03-09T20:37:11.769 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.103 metric 100 2026-03-09T20:37:11.772 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-09T20:37:11.772 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-09T20:37:11.776 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-09T20:37:11.776 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-09T20:37:11.776 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-09T20:37:11.776 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout 2: eth0: mtu 1500 state UP qlen 1000 2026-03-09T20:37:11.776 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout inet6 fe80::5055:ff:fe00:3/64 scope link noprefixroute 2026-03-09T20:37:11.776 INFO:teuthology.orchestra.run.vm03.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-09T20:37:11.776 INFO:teuthology.orchestra.run.vm03.stdout:Mon IP `192.168.123.103` is in CIDR network `192.168.123.0/24` 2026-03-09T20:37:11.776 INFO:teuthology.orchestra.run.vm03.stdout:Mon IP `192.168.123.103` is in CIDR network `192.168.123.0/24` 2026-03-09T20:37:11.776 INFO:teuthology.orchestra.run.vm03.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24'] 2026-03-09T20:37:11.777 INFO:teuthology.orchestra.run.vm03.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-09T20:37:11.777 INFO:teuthology.orchestra.run.vm03.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T20:37:13.207 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stdout 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-09T20:37:13.207 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stderr Trying to pull quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T20:37:13.207 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stderr Getting image source signatures 2026-03-09T20:37:13.207 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stderr Copying blob sha256:1752b8d01aa0dd33bbe0ab24e8316174c94fbdcd5d26252e2680bba0624747a7 2026-03-09T20:37:13.207 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stderr Copying blob sha256:8e380faede39ebd4286247457b408d979ab568aafd8389c42ec304b8cfba4e92 2026-03-09T20:37:13.207 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stderr Copying config sha256:654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-09T20:37:13.207 INFO:teuthology.orchestra.run.vm03.stdout:/bin/podman: stderr Writing manifest to image destination 2026-03-09T20:37:13.362 INFO:teuthology.orchestra.run.vm03.stdout:ceph: stdout ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-09T20:37:13.362 INFO:teuthology.orchestra.run.vm03.stdout:Ceph version: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-09T20:37:13.362 INFO:teuthology.orchestra.run.vm03.stdout:Extracting ceph user uid/gid from container image... 2026-03-09T20:37:13.839 INFO:teuthology.orchestra.run.vm03.stdout:stat: stdout 167 167 2026-03-09T20:37:13.839 INFO:teuthology.orchestra.run.vm03.stdout:Creating initial keys... 2026-03-09T20:37:13.965 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-authtool: stdout AQB5L69pe7pDNxAApaRJmJXxRfgr08ww0gWDdQ== 2026-03-09T20:37:14.068 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-authtool: stdout AQB6L69p+Bd/AhAAHYxKCFBbIq/ESfdQEGYkgA== 2026-03-09T20:37:14.176 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph-authtool: stdout AQB6L69pSgMmCBAAvMFiFui9hSVFk62S5YKr6A== 2026-03-09T20:37:14.176 INFO:teuthology.orchestra.run.vm03.stdout:Creating initial monmap... 2026-03-09T20:37:14.269 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-09T20:37:14.269 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-09T20:37:14.269 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:14.269 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-09T20:37:14.269 INFO:teuthology.orchestra.run.vm03.stdout:monmaptool for a [v2:192.168.123.103:3300,v1:192.168.123.103:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-09T20:37:14.269 INFO:teuthology.orchestra.run.vm03.stdout:setting min_mon_release = quincy 2026-03-09T20:37:14.269 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: set fsid to 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:14.269 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-09T20:37:14.269 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:37:14.269 INFO:teuthology.orchestra.run.vm03.stdout:Creating mon... 2026-03-09T20:37:14.418 INFO:teuthology.orchestra.run.vm03.stdout:create mon.a on 2026-03-09T20:37:14.587 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-09T20:37:14.724 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-09T20:37:14.863 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd.target → /etc/systemd/system/ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd.target. 2026-03-09T20:37:14.863 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd.target → /etc/systemd/system/ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd.target. 2026-03-09T20:37:15.028 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mon.a 2026-03-09T20:37:15.028 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to reset failed state of unit ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mon.a.service: Unit ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mon.a.service not loaded. 2026-03-09T20:37:15.184 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd.target.wants/ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mon.a.service → /etc/systemd/system/ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@.service. 2026-03-09T20:37:15.371 INFO:teuthology.orchestra.run.vm03.stdout:firewalld does not appear to be present 2026-03-09T20:37:15.371 INFO:teuthology.orchestra.run.vm03.stdout:Not possible to enable service . firewalld.service is not available 2026-03-09T20:37:15.371 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mon to start... 2026-03-09T20:37:15.371 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mon... 2026-03-09T20:37:15.718 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout cluster: 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout id: 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout services: 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.160375s) 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout data: 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout pgs: 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:mon is available 2026-03-09T20:37:15.719 INFO:teuthology.orchestra.run.vm03.stdout:Assimilating anything we can from ceph.conf... 2026-03-09T20:37:15.930 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T20:37:15.930 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [global] 2026-03-09T20:37:15.930 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout fsid = 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:15.930 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-09T20:37:15.930 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.103:3300,v1:192.168.123.103:6789] 2026-03-09T20:37:15.930 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-09T20:37:15.930 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-09T20:37:15.930 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-09T20:37:15.930 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-09T20:37:15.930 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T20:37:15.930 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-09T20:37:15.930 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mgr/cephadm/use_agent = True 2026-03-09T20:37:15.930 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-09T20:37:15.930 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T20:37:15.931 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [osd] 2026-03-09T20:37:15.931 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-09T20:37:15.931 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-09T20:37:15.931 INFO:teuthology.orchestra.run.vm03.stdout:Generating new minimal ceph.conf... 2026-03-09T20:37:16.119 INFO:teuthology.orchestra.run.vm03.stdout:Restarting the monitor... 2026-03-09T20:37:16.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-a[51944]: 2026-03-09T20:37:16.210+0000 7f75fc9c3640 -1 mon.a@0(leader) e1 *** Got Signal Terminated *** 2026-03-09T20:37:16.706 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 podman[52144]: 2026-03-09 20:37:16.447087921 +0000 UTC m=+0.255752183 container died 5805aca11c66b2c33c2acb0c7dd78f0f462342b5312db65fd95e229e65ef6136 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-a, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-09T20:37:16.706 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 podman[52144]: 2026-03-09 20:37:16.46145912 +0000 UTC m=+0.270123382 container remove 5805aca11c66b2c33c2acb0c7dd78f0f462342b5312db65fd95e229e65ef6136 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-a, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, CEPH_REF=squid, ceph=True, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-09T20:37:16.706 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 bash[52144]: ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-a 2026-03-09T20:37:16.706 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 systemd[1]: ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mon.a.service: Deactivated successfully. 2026-03-09T20:37:16.706 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 systemd[1]: Stopped Ceph mon.a for 8176208c-1bf7-11f1-86ba-4f626c793dbd. 2026-03-09T20:37:16.706 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 systemd[1]: Starting Ceph mon.a for 8176208c-1bf7-11f1-86ba-4f626c793dbd... 2026-03-09T20:37:16.706 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 podman[52212]: 2026-03-09 20:37:16.648693061 +0000 UTC m=+0.021366734 container create 54b853786cfb4d5a504b3b0d6aaa94f8ac8cd3b271bddadc1ad77a96e7050d30 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-a, CEPH_REF=squid, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , ceph=True, org.label-schema.build-date=20260223) 2026-03-09T20:37:16.706 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 podman[52212]: 2026-03-09 20:37:16.705266341 +0000 UTC m=+0.077940023 container init 54b853786cfb4d5a504b3b0d6aaa94f8ac8cd3b271bddadc1ad77a96e7050d30 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-a, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223) 2026-03-09T20:37:16.722 INFO:teuthology.orchestra.run.vm03.stdout:Setting public_network to 192.168.123.0/24 in mon config section 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 podman[52212]: 2026-03-09 20:37:16.709085668 +0000 UTC m=+0.081759341 container start 54b853786cfb4d5a504b3b0d6aaa94f8ac8cd3b271bddadc1ad77a96e7050d30 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-a, org.label-schema.build-date=20260223, CEPH_REF=squid, io.buildah.version=1.41.3, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 bash[52212]: 54b853786cfb4d5a504b3b0d6aaa94f8ac8cd3b271bddadc1ad77a96e7050d30 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 podman[52212]: 2026-03-09 20:37:16.64070931 +0000 UTC m=+0.013382992 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 systemd[1]: Started Ceph mon.a for 8176208c-1bf7-11f1-86ba-4f626c793dbd. 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: set uid:gid to 167:167 (ceph:ceph) 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 2 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: pidfile_write: ignore empty --pid-file 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: load: jerasure load: lrc 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: RocksDB version: 7.9.2 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Git sha 0 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Compile date 2026-02-25 18:11:04 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: DB SUMMARY 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: DB Session ID: Z373VK41Q45A4VJ90Z21 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: CURRENT file: CURRENT 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: IDENTITY file: IDENTITY 2026-03-09T20:37:17.018 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 75535 ; 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.error_if_exists: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.create_if_missing: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.paranoid_checks: 1 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.env: 0x55ec6bee2dc0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.fs: PosixFileSystem 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.info_log: 0x55ec6cadc700 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_file_opening_threads: 16 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.statistics: (nil) 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.use_fsync: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_log_file_size: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.keep_log_file_num: 1000 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.recycle_log_file_num: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.allow_fallocate: 1 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.allow_mmap_reads: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.allow_mmap_writes: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.use_direct_reads: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.create_missing_column_families: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.db_log_dir: 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.wal_dir: 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.advise_random_on_open: 1 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.db_write_buffer_size: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.write_buffer_manager: 0x55ec6cae1900 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.rate_limiter: (nil) 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.wal_recovery_mode: 2 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.enable_thread_tracking: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.enable_pipelined_write: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.unordered_write: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.row_cache: None 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.wal_filter: None 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.allow_ingest_behind: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.two_write_queues: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.manual_wal_flush: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.wal_compression: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.atomic_flush: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-09T20:37:17.019 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.log_readahead_size: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.best_efforts_recovery: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.allow_data_in_errors: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.db_host_id: __hostname__ 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_background_jobs: 2 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_background_compactions: -1 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_subcompactions: 1 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_total_wal_size: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_open_files: -1 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.bytes_per_sync: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compaction_readahead_size: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_background_flushes: -1 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Compression algorithms supported: 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: kZSTD supported: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: kXpressCompression supported: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: kBZip2Compression supported: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: kLZ4Compression supported: 1 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: kZlibCompression supported: 1 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: kLZ4HCCompression supported: 1 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: kSnappyCompression supported: 1 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.merge_operator: 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compaction_filter: None 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compaction_filter_factory: None 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.sst_partitioner_factory: None 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55ec6cadc640) 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: cache_index_and_filter_blocks: 1 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: pin_top_level_index_and_filter: 1 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: index_type: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: data_block_index_type: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: index_shortening: 1 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: checksum: 4 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: no_block_cache: 0 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: block_cache: 0x55ec6cb01350 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: block_cache_name: BinnedLRUCache 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: block_cache_options: 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: capacity : 536870912 2026-03-09T20:37:17.020 INFO:journalctl@ceph.mon.a.vm03.stdout: num_shard_bits : 4 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: strict_capacity_limit : 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: high_pri_pool_ratio: 0.000 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: block_cache_compressed: (nil) 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: persistent_cache: (nil) 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: block_size: 4096 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: block_size_deviation: 10 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: block_restart_interval: 16 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: index_block_restart_interval: 1 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: metadata_block_size: 4096 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: partition_filters: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: use_delta_encoding: 1 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: filter_policy: bloomfilter 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: whole_key_filtering: 1 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: verify_compression: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: read_amp_bytes_per_bit: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: format_version: 5 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: enable_index_compression: 1 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: block_align: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: max_auto_readahead_size: 262144 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: prepopulate_block_cache: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: initial_auto_readahead_size: 8192 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout: num_file_reads_for_auto_readahead: 2 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.write_buffer_size: 33554432 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_write_buffer_number: 2 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compression: NoCompression 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.bottommost_compression: Disabled 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.prefix_extractor: nullptr 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.num_levels: 7 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compression_opts.level: 32767 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compression_opts.strategy: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compression_opts.enabled: false 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.target_file_size_base: 67108864 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-09T20:37:17.021 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.arena_block_size: 1048576 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.disable_auto_compactions: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.inplace_update_support: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.bloom_locality: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.max_successive_merges: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.paranoid_file_checks: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.force_consistency_checks: 1 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.report_bg_io_stats: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.ttl: 2592000 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.enable_blob_files: false 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.min_blob_size: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.blob_file_size: 268435456 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.blob_file_starting_level: 0 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: d7874d8f-842d-4b97-893d-a9548f3b8e99 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773088636734850, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-09T20:37:17.022 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773088636739148, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 72616, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 225, "table_properties": {"data_size": 70895, "index_size": 174, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9705, "raw_average_key_size": 49, "raw_value_size": 65374, "raw_average_value_size": 333, "num_data_blocks": 8, "num_entries": 196, "num_filter_entries": 196, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773088636, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "d7874d8f-842d-4b97-893d-a9548f3b8e99", "db_session_id": "Z373VK41Q45A4VJ90Z21", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773088636739240, "job": 1, "event": "recovery_finished"} 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55ec6cb02e00 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: DB pointer 0x55ec6cc18000 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: ** DB Stats ** 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: ** Compaction Stats [default] ** 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: L0 2/0 72.77 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 17.1 0.00 0.00 1 0.004 0 0 0.0 0.0 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Sum 2/0 72.77 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 17.1 0.00 0.00 1 0.004 0 0 0.0 0.0 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 17.1 0.00 0.00 1 0.004 0 0 0.0 0.0 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: ** Compaction Stats [default] ** 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 17.1 0.00 0.00 1 0.004 0 0 0.0 0.0 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Cumulative compaction: 0.00 GB write, 7.59 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Interval compaction: 0.00 GB write, 7.59 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-09T20:37:17.023 INFO:journalctl@ceph.mon.a.vm03.stdout: Block cache BinnedLRUCache@0x55ec6cb01350#2 capacity: 512.00 MB usage: 1.06 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 2e-05 secs_since: 0 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout: Block cache entry stats(count,size,portion): FilterBlock(2,0.70 KB,0.00013411%) IndexBlock(2,0.36 KB,6.85453e-05%) Misc(1,0.00 KB,0%) 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout: 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: starting mon.a rank 0 at public addrs [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] at bind addrs [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: mon.a@-1(???) e1 preinit fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: monmap epoch 1 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: last_changed 2026-03-09T20:37:14.248038+0000 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: created 2026-03-09T20:37:14.248038+0000 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: min_mon_release 19 (squid) 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: election_strategy: 1 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: fsmap 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: osdmap e1: 0 total, 0 up, 0 in 2026-03-09T20:37:17.024 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:16 vm03 ceph-mon[52226]: mgrmap e1: no daemons active 2026-03-09T20:37:17.404 INFO:teuthology.orchestra.run.vm03.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-09T20:37:17.404 INFO:teuthology.orchestra.run.vm03.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:37:17.404 INFO:teuthology.orchestra.run.vm03.stdout:Creating mgr... 2026-03-09T20:37:17.405 INFO:teuthology.orchestra.run.vm03.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-09T20:37:17.405 INFO:teuthology.orchestra.run.vm03.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-09T20:37:17.571 INFO:teuthology.orchestra.run.vm03.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mgr.a 2026-03-09T20:37:17.571 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Failed to reset failed state of unit ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mgr.a.service: Unit ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mgr.a.service not loaded. 2026-03-09T20:37:17.705 INFO:teuthology.orchestra.run.vm03.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd.target.wants/ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mgr.a.service → /etc/systemd/system/ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@.service. 2026-03-09T20:37:17.907 INFO:teuthology.orchestra.run.vm03.stdout:firewalld does not appear to be present 2026-03-09T20:37:17.907 INFO:teuthology.orchestra.run.vm03.stdout:Not possible to enable service . firewalld.service is not available 2026-03-09T20:37:17.907 INFO:teuthology.orchestra.run.vm03.stdout:firewalld does not appear to be present 2026-03-09T20:37:17.907 INFO:teuthology.orchestra.run.vm03.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-09T20:37:17.907 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mgr to start... 2026-03-09T20:37:17.907 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mgr... 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsid": "8176208c-1bf7-11f1-86ba-4f626c793dbd", 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 0 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_age": 1, 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T20:37:18.156 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T20:37:15:402656+0000", 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T20:37:18.157 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:18.158 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T20:37:18.158 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:37:18.158 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T20:37:15.404689+0000", 2026-03-09T20:37:18.158 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T20:37:18.158 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:18.158 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T20:37:18.158 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-09T20:37:18.158 INFO:teuthology.orchestra.run.vm03.stdout:mgr not available, waiting (1/15)... 2026-03-09T20:37:18.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:18 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/3324809526' entity='client.admin' 2026-03-09T20:37:18.894 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:18 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:18.545+0000 7fd86c37c140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T20:37:19.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:19 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1146515697' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T20:37:19.394 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:18 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:18.901+0000 7fd86c37c140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T20:37:19.394 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:18 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T20:37:19.394 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:18 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T20:37:19.394 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:18 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: from numpy import show_config as show_numpy_config 2026-03-09T20:37:19.394 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:18 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:18.995+0000 7fd86c37c140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T20:37:19.394 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:19 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:19.040+0000 7fd86c37c140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T20:37:19.394 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:19 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:19.122+0000 7fd86c37c140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T20:37:19.943 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:19 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:19.674+0000 7fd86c37c140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T20:37:19.943 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:19 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:19.804+0000 7fd86c37c140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:37:19.943 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:19 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:19.850+0000 7fd86c37c140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T20:37:19.943 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:19 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:19.895+0000 7fd86c37c140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T20:37:19.943 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:19 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:19.942+0000 7fd86c37c140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T20:37:20.244 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:19 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:19.983+0000 7fd86c37c140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T20:37:20.245 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:20 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:20.179+0000 7fd86c37c140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T20:37:20.245 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:20 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:20.238+0000 7fd86c37c140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T20:37:20.599 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T20:37:20.599 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-09T20:37:20.599 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsid": "8176208c-1bf7-11f1-86ba-4f626c793dbd", 2026-03-09T20:37:20.599 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T20:37:20.599 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T20:37:20.599 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T20:37:20.599 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T20:37:20.599 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:20.599 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T20:37:20.599 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T20:37:20.599 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 0 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T20:37:15:402656+0000", 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T20:37:15.404689+0000", 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-09T20:37:20.600 INFO:teuthology.orchestra.run.vm03.stdout:mgr not available, waiting (2/15)... 2026-03-09T20:37:20.851 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:20 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1571332773' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T20:37:20.851 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:20 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:20.502+0000 7fd86c37c140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T20:37:21.144 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:20 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:20.850+0000 7fd86c37c140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T20:37:21.144 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:20 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:20.891+0000 7fd86c37c140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T20:37:21.144 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:20 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:20.936+0000 7fd86c37c140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T20:37:21.144 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:21.023+0000 7fd86c37c140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T20:37:21.144 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:21.066+0000 7fd86c37c140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T20:37:21.451 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:21.161+0000 7fd86c37c140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T20:37:21.451 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:21.297+0000 7fd86c37c140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:21.449+0000 7fd86c37c140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:21.495+0000 7fd86c37c140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-mon[52226]: Activating manager daemon a 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-mon[52226]: mgrmap e2: a(active, starting, since 0.00514513s) 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-mon[52226]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-mon[52226]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-mon[52226]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-mon[52226]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-mon[52226]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-mon[52226]: Manager daemon a is now available 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-mon[52226]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-mon[52226]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-mon[52226]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-mon[52226]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' 2026-03-09T20:37:21.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:21 vm03 ceph-mon[52226]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' 2026-03-09T20:37:22.926 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T20:37:22.926 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsid": "8176208c-1bf7-11f1-86ba-4f626c793dbd", 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 0 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "quorum_age": 6, 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T20:37:22.927 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T20:37:15:402656+0000", 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ], 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T20:37:15.404689+0000", 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout }, 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-09T20:37:22.928 INFO:teuthology.orchestra.run.vm03.stdout:mgr is available 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [global] 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout fsid = 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.103:3300,v1:192.168.123.103:6789] 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout [osd] 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-09T20:37:23.232 INFO:teuthology.orchestra.run.vm03.stdout:Enabling cephadm module... 2026-03-09T20:37:23.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:23 vm03 ceph-mon[52226]: mgrmap e3: a(active, since 1.01238s) 2026-03-09T20:37:23.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:23 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/3834801571' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T20:37:23.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:23 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/3424911298' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-09T20:37:23.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:23 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/3424911298' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-09T20:37:23.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:23 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1290895205' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-09T20:37:24.483 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:24 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: ignoring --setuser ceph since I am not root 2026-03-09T20:37:24.484 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:24 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: ignoring --setgroup ceph since I am not root 2026-03-09T20:37:24.484 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:24 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:24.330+0000 7fd29aa89140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T20:37:24.484 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:24 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:24.376+0000 7fd29aa89140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T20:37:24.527 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-09T20:37:24.528 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 4, 2026-03-09T20:37:24.528 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T20:37:24.528 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-09T20:37:24.528 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-09T20:37:24.528 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-09T20:37:24.528 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for the mgr to restart... 2026-03-09T20:37:24.528 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mgr epoch 4... 2026-03-09T20:37:25.144 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:24 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:24.853+0000 7fd29aa89140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T20:37:25.644 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:25 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:25.211+0000 7fd29aa89140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T20:37:25.644 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:25 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T20:37:25.644 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:25 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T20:37:25.644 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:25 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: from numpy import show_config as show_numpy_config 2026-03-09T20:37:25.644 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:25 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:25.303+0000 7fd29aa89140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T20:37:25.644 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:25 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:25.340+0000 7fd29aa89140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T20:37:25.644 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:25 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:25.416+0000 7fd29aa89140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T20:37:25.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:25 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1290895205' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-09T20:37:25.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:25 vm03 ceph-mon[52226]: mgrmap e4: a(active, since 2s) 2026-03-09T20:37:25.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:25 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1946942812' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T20:37:26.218 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:25 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:25.960+0000 7fd29aa89140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T20:37:26.218 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:26 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:26.089+0000 7fd29aa89140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:37:26.218 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:26 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:26.133+0000 7fd29aa89140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T20:37:26.218 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:26 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:26.173+0000 7fd29aa89140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T20:37:26.218 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:26 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:26.216+0000 7fd29aa89140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T20:37:26.498 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:26 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:26.255+0000 7fd29aa89140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T20:37:26.498 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:26 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:26.439+0000 7fd29aa89140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T20:37:26.894 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:26 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:26.497+0000 7fd29aa89140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T20:37:26.894 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:26 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:26.739+0000 7fd29aa89140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T20:37:27.327 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:27.033+0000 7fd29aa89140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T20:37:27.327 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:27.075+0000 7fd29aa89140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T20:37:27.327 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:27.118+0000 7fd29aa89140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T20:37:27.327 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:27.202+0000 7fd29aa89140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T20:37:27.327 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:27.241+0000 7fd29aa89140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T20:37:27.609 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:27.326+0000 7fd29aa89140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T20:37:27.609 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:27.452+0000 7fd29aa89140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:37:27.894 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:27.608+0000 7fd29aa89140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T20:37:27.894 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:27.650+0000 7fd29aa89140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: Active manager daemon a restarted 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: Activating manager daemon a 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: osdmap e2: 0 total, 0 up, 0 in 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: mgrmap e5: a(active, starting, since 0.231904s) 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: Manager daemon a is now available 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:28.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:27 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:28.951 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-09T20:37:28.951 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 6, 2026-03-09T20:37:28.951 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-09T20:37:28.951 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-09T20:37:28.952 INFO:teuthology.orchestra.run.vm03.stdout:mgr epoch 4 is available 2026-03-09T20:37:28.952 INFO:teuthology.orchestra.run.vm03.stdout:Setting orchestrator backend to cephadm... 2026-03-09T20:37:29.221 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:28 vm03 ceph-mon[52226]: Found migration_current of "None". Setting to last migration. 2026-03-09T20:37:29.221 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:28 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:37:29.221 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:28 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:37:29.221 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:28 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:29.221 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:28 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:29.221 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:28 vm03 ceph-mon[52226]: mgrmap e6: a(active, since 1.23515s) 2026-03-09T20:37:29.497 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-09T20:37:29.497 INFO:teuthology.orchestra.run.vm03.stdout:Generating ssh key... 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: Generating public/private ed25519 key pair. 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: Your identification has been saved in /tmp/tmp9putr7el/key 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: Your public key has been saved in /tmp/tmp9putr7el/key.pub 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: The key fingerprint is: 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: SHA256:2Y9BNbU9fasA3bsahG4VrnZ0oNiZuOosaHbNQ2kR9A4 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: The key's randomart image is: 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: +--[ED25519 256]--+ 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: | .. o.. | 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: | .. .... o.| 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: | E.. ..+ ...+| 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: | .o ++B o . +| 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: | o+S*o* o . | 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: | + o =+o o | 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: | . = . =.o.o | 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: | + o.+. o . o | 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: |o . o+. . | 2026-03-09T20:37:30.013 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:29 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: +----[SHA256]-----+ 2026-03-09T20:37:30.065 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIL7y9YWXfJBD2xXDCV2T1Zsj1yHn7l+5+EkRzJz2Jw1L ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:30.065 INFO:teuthology.orchestra.run.vm03.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-09T20:37:30.065 INFO:teuthology.orchestra.run.vm03.stdout:Adding key to root@localhost authorized_keys... 2026-03-09T20:37:30.065 INFO:teuthology.orchestra.run.vm03.stdout:Adding host vm03... 2026-03-09T20:37:30.320 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:30 vm03 ceph-mon[52226]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T20:37:30.320 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:30 vm03 ceph-mon[52226]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T20:37:30.320 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:30 vm03 ceph-mon[52226]: [09/Mar/2026:20:37:29] ENGINE Bus STARTING 2026-03-09T20:37:30.320 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:30 vm03 ceph-mon[52226]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:30.320 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:30 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:30.320 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:30 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:30.320 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:30 vm03 ceph-mon[52226]: [09/Mar/2026:20:37:29] ENGINE Serving on http://192.168.123.103:8765 2026-03-09T20:37:30.320 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:30 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:30.320 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:30 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:30.320 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:30 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:31 vm03 ceph-mon[52226]: [09/Mar/2026:20:37:29] ENGINE Serving on https://192.168.123.103:7150 2026-03-09T20:37:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:31 vm03 ceph-mon[52226]: [09/Mar/2026:20:37:29] ENGINE Bus STARTED 2026-03-09T20:37:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:31 vm03 ceph-mon[52226]: [09/Mar/2026:20:37:29] ENGINE Client ('192.168.123.103', 35350) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T20:37:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:31 vm03 ceph-mon[52226]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:31 vm03 ceph-mon[52226]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:31 vm03 ceph-mon[52226]: Generating ssh key... 2026-03-09T20:37:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:31 vm03 ceph-mon[52226]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:31.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:31 vm03 ceph-mon[52226]: mgrmap e7: a(active, since 2s) 2026-03-09T20:37:31.940 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Added host 'vm03' with addr '192.168.123.103' 2026-03-09T20:37:31.941 INFO:teuthology.orchestra.run.vm03.stdout:Deploying unmanaged mon service... 2026-03-09T20:37:32.231 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-09T20:37:32.231 INFO:teuthology.orchestra.run.vm03.stdout:Deploying unmanaged mgr service... 2026-03-09T20:37:32.473 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:32 vm03 ceph-mon[52226]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm03", "addr": "192.168.123.103", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:32.473 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:32 vm03 ceph-mon[52226]: Deploying cephadm binary to vm03 2026-03-09T20:37:32.473 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:32 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:32.473 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:32 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:32.473 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:32 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:32.524 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-09T20:37:33.179 INFO:teuthology.orchestra.run.vm03.stdout:Enabling the dashboard module... 2026-03-09T20:37:33.392 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:33 vm03 ceph-mon[52226]: Added host vm03 2026-03-09T20:37:33.392 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:33 vm03 ceph-mon[52226]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:33.392 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:33 vm03 ceph-mon[52226]: Saving service mon spec with placement count:5 2026-03-09T20:37:33.392 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:33 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:33.392 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:33 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1373298368' entity='client.admin' 2026-03-09T20:37:33.392 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:33 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/3194674838' entity='client.admin' 2026-03-09T20:37:34.264 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-mon[52226]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:34.264 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-mon[52226]: Saving service mgr spec with placement count:2 2026-03-09T20:37:34.264 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/352927836' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-09T20:37:34.264 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:34.264 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:34.264 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:34.264 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:34.264 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:37:34.264 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:34.264 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-09T20:37:34.264 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-mon[52226]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]': finished 2026-03-09T20:37:34.590 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: ignoring --setuser ceph since I am not root 2026-03-09T20:37:34.590 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: ignoring --setgroup ceph since I am not root 2026-03-09T20:37:34.590 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:34.417+0000 7fd3b41e5140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T20:37:34.590 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:34.468+0000 7fd3b41e5140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T20:37:34.622 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-09T20:37:34.622 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "epoch": 8, 2026-03-09T20:37:34.622 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T20:37:34.622 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-09T20:37:34.622 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-09T20:37:34.622 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-09T20:37:34.622 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for the mgr to restart... 2026-03-09T20:37:34.622 INFO:teuthology.orchestra.run.vm03.stdout:Waiting for mgr epoch 8... 2026-03-09T20:37:35.221 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:34 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:34.935+0000 7fd3b41e5140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T20:37:35.499 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:35 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/352927836' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-09T20:37:35.499 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:35 vm03 ceph-mon[52226]: mgrmap e8: a(active, since 6s) 2026-03-09T20:37:35.499 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:35 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/195646652' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T20:37:35.499 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:35 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:35.287+0000 7fd3b41e5140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T20:37:35.499 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:35 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T20:37:35.499 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:35 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T20:37:35.499 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:35 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: from numpy import show_config as show_numpy_config 2026-03-09T20:37:35.499 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:35 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:35.379+0000 7fd3b41e5140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T20:37:35.499 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:35 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:35.419+0000 7fd3b41e5140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T20:37:35.499 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:35 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:35.497+0000 7fd3b41e5140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T20:37:36.311 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:36 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:36.028+0000 7fd3b41e5140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T20:37:36.311 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:36 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:36.146+0000 7fd3b41e5140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:37:36.311 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:36 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:36.193+0000 7fd3b41e5140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T20:37:36.312 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:36 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:36.230+0000 7fd3b41e5140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T20:37:36.312 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:36 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:36.272+0000 7fd3b41e5140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T20:37:36.644 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:36 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:36.310+0000 7fd3b41e5140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T20:37:36.644 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:36 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:36.494+0000 7fd3b41e5140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T20:37:36.644 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:36 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:36.552+0000 7fd3b41e5140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T20:37:37.079 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:36 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:36.787+0000 7fd3b41e5140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T20:37:37.364 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:37.078+0000 7fd3b41e5140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T20:37:37.364 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:37.118+0000 7fd3b41e5140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T20:37:37.364 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:37.162+0000 7fd3b41e5140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T20:37:37.364 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:37.243+0000 7fd3b41e5140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T20:37:37.364 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:37.281+0000 7fd3b41e5140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T20:37:37.619 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:37.363+0000 7fd3b41e5140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T20:37:37.619 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:37.476+0000 7fd3b41e5140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:37:37.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-mon[52226]: Active manager daemon a restarted 2026-03-09T20:37:37.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-mon[52226]: Activating manager daemon a 2026-03-09T20:37:37.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-mon[52226]: osdmap e3: 0 total, 0 up, 0 in 2026-03-09T20:37:37.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-mon[52226]: mgrmap e9: a(active, starting, since 0.00536758s) 2026-03-09T20:37:37.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:37:37.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:37:37.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:37:37.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:37:37.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:37:37.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-mon[52226]: Manager daemon a is now available 2026-03-09T20:37:37.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:37.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:37.895 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:37.618+0000 7fd3b41e5140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T20:37:37.895 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:37:37 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:37:37.654+0000 7fd3b41e5140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T20:37:38.728 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout { 2026-03-09T20:37:38.729 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 10, 2026-03-09T20:37:38.729 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-09T20:37:38.729 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout } 2026-03-09T20:37:38.729 INFO:teuthology.orchestra.run.vm03.stdout:mgr epoch 8 is available 2026-03-09T20:37:38.729 INFO:teuthology.orchestra.run.vm03.stdout:Generating a dashboard self-signed certificate... 2026-03-09T20:37:39.115 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:38 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:37:39.115 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:38 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:37:39.115 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:38 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:39.115 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:38 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:37:39.115 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:38 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:39.116 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:38 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-09T20:37:39.116 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:38 vm03 ceph-mon[52226]: mgrmap e10: a(active, since 1.0097s) 2026-03-09T20:37:39.149 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-09T20:37:39.149 INFO:teuthology.orchestra.run.vm03.stdout:Creating initial admin user... 2026-03-09T20:37:39.577 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$uVMXAswSaN7BnvLywtmD6eSOb5YEqBdTysl3QlSCprQas8bfEppCu", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773088659, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-09T20:37:39.577 INFO:teuthology.orchestra.run.vm03.stdout:Fetching dashboard port number... 2026-03-09T20:37:39.838 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stdout 8443 2026-03-09T20:37:39.838 INFO:teuthology.orchestra.run.vm03.stdout:firewalld does not appear to be present 2026-03-09T20:37:39.838 INFO:teuthology.orchestra.run.vm03.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-09T20:37:39.842 INFO:teuthology.orchestra.run.vm03.stdout:Ceph Dashboard is now available at: 2026-03-09T20:37:39.842 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:37:39.842 INFO:teuthology.orchestra.run.vm03.stdout: URL: https://vm03.local:8443/ 2026-03-09T20:37:39.842 INFO:teuthology.orchestra.run.vm03.stdout: User: admin 2026-03-09T20:37:39.842 INFO:teuthology.orchestra.run.vm03.stdout: Password: f8f6z5x5vd 2026-03-09T20:37:39.842 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:37:39.842 INFO:teuthology.orchestra.run.vm03.stdout:Saving cluster configuration to /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config directory 2026-03-09T20:37:40.120 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:40 vm03 ceph-mon[52226]: [09/Mar/2026:20:37:38] ENGINE Bus STARTING 2026-03-09T20:37:40.121 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:40 vm03 ceph-mon[52226]: [09/Mar/2026:20:37:38] ENGINE Serving on https://192.168.123.103:7150 2026-03-09T20:37:40.121 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:40 vm03 ceph-mon[52226]: [09/Mar/2026:20:37:38] ENGINE Client ('192.168.123.103', 35366) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T20:37:40.121 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:40 vm03 ceph-mon[52226]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T20:37:40.121 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:40 vm03 ceph-mon[52226]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T20:37:40.121 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:40 vm03 ceph-mon[52226]: [09/Mar/2026:20:37:38] ENGINE Serving on http://192.168.123.103:8765 2026-03-09T20:37:40.121 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:40 vm03 ceph-mon[52226]: [09/Mar/2026:20:37:38] ENGINE Bus STARTED 2026-03-09T20:37:40.121 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:40 vm03 ceph-mon[52226]: Deploying daemon agent.vm03 on vm03 2026-03-09T20:37:40.121 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:40 vm03 ceph-mon[52226]: from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:40.121 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:40 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:40.121 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:40 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:40.121 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:40 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:40.121 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:40 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/3271032368' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout:Or, if you are only running a single cluster on this host: 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout: ceph telemetry on 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout:For more information see: 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:37:40.180 INFO:teuthology.orchestra.run.vm03.stdout:Bootstrap complete. 2026-03-09T20:37:40.211 INFO:tasks.cephadm:Fetching config... 2026-03-09T20:37:40.211 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T20:37:40.212 DEBUG:teuthology.orchestra.run.vm03:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-09T20:37:40.233 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-09T20:37:40.233 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T20:37:40.233 DEBUG:teuthology.orchestra.run.vm03:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-09T20:37:40.291 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-09T20:37:40.291 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T20:37:40.291 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/keyring of=/dev/stdout 2026-03-09T20:37:40.437 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-09T20:37:40.437 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T20:37:40.437 DEBUG:teuthology.orchestra.run.vm03:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-09T20:37:40.500 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-09T20:37:40.500 DEBUG:teuthology.orchestra.run.vm03:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIL7y9YWXfJBD2xXDCV2T1Zsj1yHn7l+5+EkRzJz2Jw1L ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T20:37:40.616 INFO:teuthology.orchestra.run.vm03.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIL7y9YWXfJBD2xXDCV2T1Zsj1yHn7l+5+EkRzJz2Jw1L ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:40.624 DEBUG:teuthology.orchestra.run.vm04:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIL7y9YWXfJBD2xXDCV2T1Zsj1yHn7l+5+EkRzJz2Jw1L ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T20:37:40.661 INFO:teuthology.orchestra.run.vm04.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIL7y9YWXfJBD2xXDCV2T1Zsj1yHn7l+5+EkRzJz2Jw1L ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:40.673 DEBUG:teuthology.orchestra.run.vm08:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIL7y9YWXfJBD2xXDCV2T1Zsj1yHn7l+5+EkRzJz2Jw1L ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T20:37:40.711 INFO:teuthology.orchestra.run.vm08.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIL7y9YWXfJBD2xXDCV2T1Zsj1yHn7l+5+EkRzJz2Jw1L ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:40.722 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-09T20:37:41.042 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:37:41.373 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:41 vm03 ceph-mon[52226]: from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:41.374 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:41 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/3807724702' entity='client.admin' 2026-03-09T20:37:41.374 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:41 vm03 ceph-mon[52226]: mgrmap e11: a(active, since 3s) 2026-03-09T20:37:41.374 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:41.374 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:41.374 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:41.374 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:41.374 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:41.374 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:41.374 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:41.374 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:41.374 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:41.374 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:41.374 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:41.499 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-09T20:37:41.499 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-09T20:37:41.819 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:37:42.162 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm04 2026-03-09T20:37:42.162 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T20:37:42.162 DEBUG:teuthology.orchestra.run.vm04:> dd of=/etc/ceph/ceph.conf 2026-03-09T20:37:42.178 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T20:37:42.178 DEBUG:teuthology.orchestra.run.vm04:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:37:42.234 INFO:tasks.cephadm:Adding host vm04 to orchestrator... 2026-03-09T20:37:42.234 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph orch host add vm04 2026-03-09T20:37:42.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:42.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:42 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1393567455' entity='client.admin' 2026-03-09T20:37:42.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:42.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:42.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:42.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:42.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:42.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:42.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:42.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:42.502 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:37:43.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:43 vm03 ceph-mon[52226]: from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:43.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:43 vm03 ceph-mon[52226]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T20:37:43.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:43 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:43.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:43 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:43.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:43 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:43.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:43 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:43.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:43 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:43.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:43 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:44.353 INFO:teuthology.orchestra.run.vm03.stdout:Added host 'vm04' with addr '192.168.123.104' 2026-03-09T20:37:44.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:44 vm03 ceph-mon[52226]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:44.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:44 vm03 ceph-mon[52226]: Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:37:44.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:44 vm03 ceph-mon[52226]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm04", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:44.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:44 vm03 ceph-mon[52226]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:37:44.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:44 vm03 ceph-mon[52226]: Deploying cephadm binary to vm04 2026-03-09T20:37:44.409 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph orch host ls --format=json 2026-03-09T20:37:44.586 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:37:44.825 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:37:44.825 INFO:teuthology.orchestra.run.vm03.stdout:[{"addr": "192.168.123.103", "hostname": "vm03", "labels": [], "status": ""}, {"addr": "192.168.123.104", "hostname": "vm04", "labels": [], "status": ""}] 2026-03-09T20:37:44.897 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm08 2026-03-09T20:37:44.897 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:37:44.897 DEBUG:teuthology.orchestra.run.vm08:> dd of=/etc/ceph/ceph.conf 2026-03-09T20:37:44.913 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:37:44.913 DEBUG:teuthology.orchestra.run.vm08:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:37:44.972 INFO:tasks.cephadm:Adding host vm08 to orchestrator... 2026-03-09T20:37:44.972 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph orch host add vm08 2026-03-09T20:37:45.155 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:37:45.272 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:45 vm03 ceph-mon[52226]: mgrmap e12: a(active, since 6s) 2026-03-09T20:37:45.273 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:45 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:45.273 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:45 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:45.273 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:45 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:45.273 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:45 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:45.273 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:45 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:45.273 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:45 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:37:45.273 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:45 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:45.273 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:45 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:46.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:46 vm03 ceph-mon[52226]: Added host vm04 2026-03-09T20:37:46.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:46 vm03 ceph-mon[52226]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T20:37:46.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:46 vm03 ceph-mon[52226]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:37:46.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:46 vm03 ceph-mon[52226]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:46.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:46 vm03 ceph-mon[52226]: Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:37:46.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:46 vm03 ceph-mon[52226]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:37:46.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:46 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:46.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:46 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:46.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:46 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm04", "caps": []}]: dispatch 2026-03-09T20:37:46.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:46 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm04", "caps": []}]': finished 2026-03-09T20:37:46.974 INFO:teuthology.orchestra.run.vm03.stdout:Added host 'vm08' with addr '192.168.123.108' 2026-03-09T20:37:47.024 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph orch host ls --format=json 2026-03-09T20:37:47.221 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:37:47.239 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm08", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:47.239 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: Deploying cephadm binary to vm08 2026-03-09T20:37:47.239 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: Deploying daemon agent.vm04 on vm04 2026-03-09T20:37:47.469 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:37:47.469 INFO:teuthology.orchestra.run.vm03.stdout:[{"addr": "192.168.123.103", "hostname": "vm03", "labels": [], "status": ""}, {"addr": "192.168.123.104", "hostname": "vm04", "labels": [], "status": ""}, {"addr": "192.168.123.108", "hostname": "vm08", "labels": [], "status": ""}] 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:47.507 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:47.549 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-09T20:37:47.549 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph osd crush tunables default 2026-03-09T20:37:47.736 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:37:48.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:48 vm03 ceph-mon[52226]: Added host vm08 2026-03-09T20:37:48.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:48.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:48.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:37:48.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:48.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:48.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:48.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:48.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:48.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm08", "caps": []}]: dispatch 2026-03-09T20:37:48.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm08", "caps": []}]': finished 2026-03-09T20:37:48.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:48 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1248661400' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-09T20:37:48.896 INFO:teuthology.orchestra.run.vm03.stderr:adjusted tunables profile to default 2026-03-09T20:37:48.945 INFO:tasks.cephadm:Adding mon.a on vm03 2026-03-09T20:37:48.945 INFO:tasks.cephadm:Adding mon.b on vm04 2026-03-09T20:37:48.945 INFO:tasks.cephadm:Adding mon.c on vm08 2026-03-09T20:37:48.945 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph orch apply mon '3;vm03:192.168.123.103=a;vm04:192.168.123.104=b;vm08:192.168.123.108=c' 2026-03-09T20:37:49.191 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: Deploying daemon agent.vm08 on vm08 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1248661400' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:49.253 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:49 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:49.483 INFO:teuthology.orchestra.run.vm08.stdout:Scheduled mon update... 2026-03-09T20:37:49.540 DEBUG:teuthology.orchestra.run.vm04:mon.b> sudo journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mon.b.service 2026-03-09T20:37:49.542 DEBUG:teuthology.orchestra.run.vm08:mon.c> sudo journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mon.c.service 2026-03-09T20:37:49.544 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-09T20:37:49.544 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph mon dump -f json 2026-03-09T20:37:49.803 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:50.175 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:37:50.175 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":1,"fsid":"8176208c-1bf7-11f1-86ba-4f626c793dbd","modified":"2026-03-09T20:37:14.248038Z","created":"2026-03-09T20:37:14.248038Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:3300","nonce":0},{"type":"v1","addr":"192.168.123.103:6789","nonce":0}]},"addr":"192.168.123.103:6789/0","public_addr":"192.168.123.103:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T20:37:50.175 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 1 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm03:192.168.123.103=a;vm04:192.168.123.104=b;vm08:192.168.123.108=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: Saving service mon spec with placement vm03:192.168.123.103=a;vm04:192.168.123.104=b;vm08:192.168.123.108=c;count:3 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: Filtered out host vm08: does not belong to mon public_network(s): 192.168.123.0/24, host network(s): 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: Deploying daemon mon.b on vm04 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: from='client.? 192.168.123.108:0/436973558' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:50.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:50 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:50.898 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 systemd[1]: Started Ceph mon.b for 8176208c-1bf7-11f1-86ba-4f626c793dbd. 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: set uid:gid to 167:167 (ceph:ceph) 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 2 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: pidfile_write: ignore empty --pid-file 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: load: jerasure load: lrc 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: RocksDB version: 7.9.2 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Git sha 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Compile date 2026-02-25 18:11:04 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: DB SUMMARY 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: DB Session ID: H0A6GYNYT5GX6OQXS0PZ 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: CURRENT file: CURRENT 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: IDENTITY file: IDENTITY 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: MANIFEST file: MANIFEST-000005 size: 59 Bytes 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: SST files in /var/lib/ceph/mon/ceph-b/store.db dir, Total Num: 0, files: 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-b/store.db: 000004.log size: 511 ; 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.error_if_exists: 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.create_if_missing: 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.paranoid_checks: 1 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.env: 0x560c95a4ddc0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.fs: PosixFileSystem 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.info_log: 0x560c97c1e700 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_file_opening_threads: 16 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.statistics: (nil) 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.use_fsync: 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_log_file_size: 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.keep_log_file_num: 1000 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.recycle_log_file_num: 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.allow_fallocate: 1 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.allow_mmap_reads: 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.allow_mmap_writes: 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.use_direct_reads: 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.create_missing_column_families: 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.db_log_dir: 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.wal_dir: 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-09T20:37:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.advise_random_on_open: 1 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.db_write_buffer_size: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.write_buffer_manager: 0x560c97c23900 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.rate_limiter: (nil) 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.wal_recovery_mode: 2 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.enable_thread_tracking: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.enable_pipelined_write: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.unordered_write: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.row_cache: None 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.wal_filter: None 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.allow_ingest_behind: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.two_write_queues: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.manual_wal_flush: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.wal_compression: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.atomic_flush: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.log_readahead_size: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.best_efforts_recovery: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.allow_data_in_errors: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.db_host_id: __hostname__ 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_background_jobs: 2 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_background_compactions: -1 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_subcompactions: 1 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_total_wal_size: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_open_files: -1 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.bytes_per_sync: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compaction_readahead_size: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_background_flushes: -1 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Compression algorithms supported: 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: kZSTD supported: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: kXpressCompression supported: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: kBZip2Compression supported: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: kLZ4Compression supported: 1 2026-03-09T20:37:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: kZlibCompression supported: 1 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: kLZ4HCCompression supported: 1 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: kSnappyCompression supported: 1 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000005 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.merge_operator: 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compaction_filter: None 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compaction_filter_factory: None 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.sst_partitioner_factory: None 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x560c97c1e640) 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: cache_index_and_filter_blocks: 1 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: pin_top_level_index_and_filter: 1 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: index_type: 0 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: data_block_index_type: 0 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: index_shortening: 1 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: checksum: 4 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: no_block_cache: 0 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: block_cache: 0x560c97c43350 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: block_cache_name: BinnedLRUCache 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: block_cache_options: 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: capacity : 536870912 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: num_shard_bits : 4 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: strict_capacity_limit : 0 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: high_pri_pool_ratio: 0.000 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: block_cache_compressed: (nil) 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: persistent_cache: (nil) 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: block_size: 4096 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: block_size_deviation: 10 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: block_restart_interval: 16 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: index_block_restart_interval: 1 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: metadata_block_size: 4096 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: partition_filters: 0 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: use_delta_encoding: 1 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: filter_policy: bloomfilter 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: whole_key_filtering: 1 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: verify_compression: 0 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: read_amp_bytes_per_bit: 0 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: format_version: 5 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: enable_index_compression: 1 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: block_align: 0 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: max_auto_readahead_size: 262144 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: prepopulate_block_cache: 0 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: initial_auto_readahead_size: 8192 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout: num_file_reads_for_auto_readahead: 2 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.write_buffer_size: 33554432 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_write_buffer_number: 2 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compression: NoCompression 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.bottommost_compression: Disabled 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.prefix_extractor: nullptr 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.num_levels: 7 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-09T20:37:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compression_opts.level: 32767 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compression_opts.strategy: 0 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-09T20:37:51.261 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-09T20:37:51.261 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph mon dump -f json 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-09T20:37:51.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compression_opts.enabled: false 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.target_file_size_base: 67108864 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.arena_block_size: 1048576 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.disable_auto_compactions: 0 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.inplace_update_support: 0 2026-03-09T20:37:51.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.bloom_locality: 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.max_successive_merges: 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.paranoid_file_checks: 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.force_consistency_checks: 1 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.report_bg_io_stats: 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.ttl: 2592000 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.enable_blob_files: false 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.min_blob_size: 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.blob_file_size: 268435456 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.blob_file_starting_level: 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000005 succeeded,manifest_file_number is 5, next_file_number is 7, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 775621e7-67cf-4a86-b7f4-fec8f965b5b9 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773088670921429, "job": 1, "event": "recovery_started", "wal_files": [4]} 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #4 mode 2 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773088670922148, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1643, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 1, "largest_seqno": 5, "table_properties": {"data_size": 523, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 401, "raw_average_value_size": 80, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773088670, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "775621e7-67cf-4a86-b7f4-fec8f965b5b9", "db_session_id": "H0A6GYNYT5GX6OQXS0PZ", "orig_file_number": 8, "seqno_to_time_mapping": "N/A"}} 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773088670922220, "job": 1, "event": "recovery_finished"} 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: [db/version_set.cc:5047] Creating manifest 10 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x560c97c44e00 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: DB pointer 0x560c97d5a000 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: ** DB Stats ** 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: ** Compaction Stats [default] ** 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: L0 1/0 1.60 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 2.2 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: Sum 1/0 1.60 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 2.2 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 2.2 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: ** Compaction Stats [default] ** 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-09T20:37:51.263 INFO:journalctl@ceph.mon.b.vm04.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 2.2 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: Cumulative compaction: 0.00 GB write, 0.17 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: Interval compaction: 0.00 GB write, 0.17 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: Block cache BinnedLRUCache@0x560c97c43350#2 capacity: 512.00 MB usage: 0.22 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 9e-06 secs_since: 0 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: Block cache entry stats(count,size,portion): FilterBlock(1,0.11 KB,2.08616e-05%) IndexBlock(1,0.11 KB,2.08616e-05%) Misc(1,0.00 KB,0%) 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b does not exist in monmap, will attempt to join an existing cluster 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: using public_addr v2:192.168.123.104:0/0 -> [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: starting mon.b rank -1 at public addrs [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] at bind addrs [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon_data /var/lib/ceph/mon/ceph-b fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b@-1(???) e0 preinit fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b@-1(synchronizing).mds e1 new map 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b@-1(synchronizing).mds e1 print_map 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: e1 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: btime 2026-03-09T20:37:15:402656+0000 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: legacy client fscid: -1 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout: No filesystems configured 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b@-1(synchronizing).osd e4 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mkfs 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: monmap epoch 1 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: last_changed 2026-03-09T20:37:14.248038+0000 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: created 2026-03-09T20:37:14.248038+0000 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: min_mon_release 19 (squid) 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: election_strategy: 1 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: fsmap 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: osdmap e1: 0 total, 0 up, 0 in 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mgrmap e1: no daemons active 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/3324809526' entity='client.admin' 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1146515697' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1571332773' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Activating manager daemon a 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mgrmap e2: a(active, starting, since 0.00514513s) 2026-03-09T20:37:51.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Manager daemon a is now available 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mgrmap e3: a(active, since 1.01238s) 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/3834801571' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/3424911298' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/3424911298' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1290895205' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1290895205' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mgrmap e4: a(active, since 2s) 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1946942812' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Active manager daemon a restarted 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Activating manager daemon a 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: osdmap e2: 0 total, 0 up, 0 in 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mgrmap e5: a(active, starting, since 0.231904s) 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Manager daemon a is now available 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Found migration_current of "None". Setting to last migration. 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mgrmap e6: a(active, since 1.23515s) 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: [09/Mar/2026:20:37:29] ENGINE Bus STARTING 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: [09/Mar/2026:20:37:29] ENGINE Serving on http://192.168.123.103:8765 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: [09/Mar/2026:20:37:29] ENGINE Serving on https://192.168.123.103:7150 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: [09/Mar/2026:20:37:29] ENGINE Bus STARTED 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: [09/Mar/2026:20:37:29] ENGINE Client ('192.168.123.103', 35350) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Generating ssh key... 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mgrmap e7: a(active, since 2s) 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm03", "addr": "192.168.123.103", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Deploying cephadm binary to vm03 2026-03-09T20:37:51.265 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Added host vm03 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Saving service mon spec with placement count:5 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1373298368' entity='client.admin' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/3194674838' entity='client.admin' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Saving service mgr spec with placement count:2 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/352927836' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]': finished 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/352927836' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mgrmap e8: a(active, since 6s) 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/195646652' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Active manager daemon a restarted 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Activating manager daemon a 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: osdmap e3: 0 total, 0 up, 0 in 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mgrmap e9: a(active, starting, since 0.00536758s) 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Manager daemon a is now available 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mgrmap e10: a(active, since 1.0097s) 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: [09/Mar/2026:20:37:38] ENGINE Bus STARTING 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: [09/Mar/2026:20:37:38] ENGINE Serving on https://192.168.123.103:7150 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: [09/Mar/2026:20:37:38] ENGINE Client ('192.168.123.103', 35366) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: [09/Mar/2026:20:37:38] ENGINE Serving on http://192.168.123.103:8765 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: [09/Mar/2026:20:37:38] ENGINE Bus STARTED 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Deploying daemon agent.vm03 on vm03 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/3271032368' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-09T20:37:51.266 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/3807724702' entity='client.admin' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mgrmap e11: a(active, since 3s) 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1393567455' entity='client.admin' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm04", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Deploying cephadm binary to vm04 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mgrmap e12: a(active, since 6s) 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Added host vm04 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm04", "caps": []}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm04", "caps": []}]': finished 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm08", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Deploying cephadm binary to vm08 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Deploying daemon agent.vm04 on vm04 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.267 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Added host vm08 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm08", "caps": []}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm08", "caps": []}]': finished 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1248661400' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Deploying daemon agent.vm08 on vm08 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1248661400' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm03:192.168.123.103=a;vm04:192.168.123.104=b;vm08:192.168.123.108=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Saving service mon spec with placement vm03:192.168.123.103=a;vm04:192.168.123.104=b;vm08:192.168.123.108=c;count:3 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Filtered out host vm08: does not belong to mon public_network(s): 192.168.123.0/24, host network(s): 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: Deploying daemon mon.b on vm04 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.108:0/436973558' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.268 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:51.269 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:50 vm04 ceph-mon[53250]: mon.b@-1(synchronizing).paxosservice(auth 1..6) refresh upgraded, format 0 -> 3 2026-03-09T20:37:51.456 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: mon.a calling monitor election 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: mon.b calling monitor election 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: monmap epoch 2 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: last_changed 2026-03-09T20:37:50.953583+0000 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: created 2026-03-09T20:37:14.248038+0000 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: min_mon_release 19 (squid) 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: election_strategy: 1 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: 1: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: fsmap 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: mgrmap e12: a(active, since 18s) 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: overall HEALTH_OK 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:56.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:56.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:56.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:56.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:56.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:56.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:56.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:56 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: mon.a calling monitor election 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: mon.b calling monitor election 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: monmap epoch 2 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: last_changed 2026-03-09T20:37:50.953583+0000 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: created 2026-03-09T20:37:14.248038+0000 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: min_mon_release 19 (squid) 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: election_strategy: 1 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: 1: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: fsmap 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: mgrmap e12: a(active, since 18s) 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: overall HEALTH_OK 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:56.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:56.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:56.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:56 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:57.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:37:57 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:57.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:37:57 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:57.743 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:37:57.743 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":2,"fsid":"8176208c-1bf7-11f1-86ba-4f626c793dbd","modified":"2026-03-09T20:37:50.953583Z","created":"2026-03-09T20:37:14.248038Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:3300","nonce":0},{"type":"v1","addr":"192.168.123.103:6789","nonce":0}]},"addr":"192.168.123.103:6789/0","public_addr":"192.168.123.103:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:3300","nonce":0},{"type":"v1","addr":"192.168.123.104:6789","nonce":0}]},"addr":"192.168.123.104:6789/0","public_addr":"192.168.123.104:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-09T20:37:57.743 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 2 2026-03-09T20:37:57.812 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 podman[54453]: 2026-03-09 20:37:57.806045689 +0000 UTC m=+0.087469425 container start e84d0e3a654bffec6de3c9ae2833510d0ccae7040985a551b2c67b519f261547 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-c, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, CEPH_REF=squid, org.label-schema.build-date=20260223, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-09T20:37:58.071 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 bash[54453]: e84d0e3a654bffec6de3c9ae2833510d0ccae7040985a551b2c67b519f261547 2026-03-09T20:37:58.071 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 podman[54453]: 2026-03-09 20:37:57.731224267 +0000 UTC m=+0.012648003 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T20:37:58.071 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 systemd[1]: Started Ceph mon.c for 8176208c-1bf7-11f1-86ba-4f626c793dbd. 2026-03-09T20:37:58.071 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: set uid:gid to 167:167 (ceph:ceph) 2026-03-09T20:37:58.071 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 2 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: pidfile_write: ignore empty --pid-file 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: load: jerasure load: lrc 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: RocksDB version: 7.9.2 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Git sha 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Compile date 2026-02-25 18:11:04 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: DB SUMMARY 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: DB Session ID: JRI664OFBPNTT9XIQEES 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: CURRENT file: CURRENT 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: IDENTITY file: IDENTITY 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: MANIFEST file: MANIFEST-000005 size: 59 Bytes 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: SST files in /var/lib/ceph/mon/ceph-c/store.db dir, Total Num: 0, files: 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-c/store.db: 000004.log size: 636 ; 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.error_if_exists: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.create_if_missing: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.paranoid_checks: 1 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.env: 0x55614f29bdc0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.fs: PosixFileSystem 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.info_log: 0x55614fb23880 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_file_opening_threads: 16 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.statistics: (nil) 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.use_fsync: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_log_file_size: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.keep_log_file_num: 1000 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.recycle_log_file_num: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.allow_fallocate: 1 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.allow_mmap_reads: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.allow_mmap_writes: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.use_direct_reads: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.create_missing_column_families: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.db_log_dir: 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.wal_dir: 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.advise_random_on_open: 1 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.db_write_buffer_size: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.write_buffer_manager: 0x55614fb27900 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.rate_limiter: (nil) 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.wal_recovery_mode: 2 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.enable_thread_tracking: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.enable_pipelined_write: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.unordered_write: 0 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.row_cache: None 2026-03-09T20:37:58.072 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.wal_filter: None 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.allow_ingest_behind: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.two_write_queues: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.manual_wal_flush: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.wal_compression: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.atomic_flush: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.log_readahead_size: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.best_efforts_recovery: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.allow_data_in_errors: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.db_host_id: __hostname__ 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_background_jobs: 2 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_background_compactions: -1 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_subcompactions: 1 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_total_wal_size: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_open_files: -1 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.bytes_per_sync: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compaction_readahead_size: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_background_flushes: -1 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Compression algorithms supported: 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: kZSTD supported: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: kXpressCompression supported: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: kBZip2Compression supported: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: kLZ4Compression supported: 1 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: kZlibCompression supported: 1 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: kLZ4HCCompression supported: 1 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: kSnappyCompression supported: 1 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-c/store.db/MANIFEST-000005 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.merge_operator: 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compaction_filter: None 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compaction_filter_factory: None 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.sst_partitioner_factory: None 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55614fb234e0) 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout: cache_index_and_filter_blocks: 1 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout: pin_top_level_index_and_filter: 1 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout: index_type: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout: data_block_index_type: 0 2026-03-09T20:37:58.073 INFO:journalctl@ceph.mon.c.vm08.stdout: index_shortening: 1 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: checksum: 4 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: no_block_cache: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: block_cache: 0x55614fb469b0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: block_cache_name: BinnedLRUCache 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: block_cache_options: 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: capacity : 536870912 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: num_shard_bits : 4 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: strict_capacity_limit : 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: high_pri_pool_ratio: 0.000 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: block_cache_compressed: (nil) 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: persistent_cache: (nil) 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: block_size: 4096 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: block_size_deviation: 10 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: block_restart_interval: 16 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: index_block_restart_interval: 1 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: metadata_block_size: 4096 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: partition_filters: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: use_delta_encoding: 1 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: filter_policy: bloomfilter 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: whole_key_filtering: 1 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: verify_compression: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: read_amp_bytes_per_bit: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: format_version: 5 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: enable_index_compression: 1 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: block_align: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: max_auto_readahead_size: 262144 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: prepopulate_block_cache: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: initial_auto_readahead_size: 8192 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout: num_file_reads_for_auto_readahead: 2 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.write_buffer_size: 33554432 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_write_buffer_number: 2 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compression: NoCompression 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.bottommost_compression: Disabled 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.prefix_extractor: nullptr 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.num_levels: 7 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compression_opts.level: 32767 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compression_opts.strategy: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-09T20:37:58.074 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compression_opts.enabled: false 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.target_file_size_base: 67108864 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.arena_block_size: 1048576 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.disable_auto_compactions: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.inplace_update_support: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.bloom_locality: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.max_successive_merges: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.paranoid_file_checks: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.force_consistency_checks: 1 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.report_bg_io_stats: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.ttl: 2592000 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.enable_blob_files: false 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.min_blob_size: 0 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.blob_file_size: 268435456 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-09T20:37:58.075 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.blob_file_starting_level: 0 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-c/store.db/MANIFEST-000005 succeeded,manifest_file_number is 5, next_file_number is 7, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 5becb1da-3993-4559-b42f-f06fe2c4fc29 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773088677845412, "job": 1, "event": "recovery_started", "wal_files": [4]} 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #4 mode 2 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773088677845964, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1768, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 1, "largest_seqno": 5, "table_properties": {"data_size": 648, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 526, "raw_average_value_size": 105, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773088677, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "5becb1da-3993-4559-b42f-f06fe2c4fc29", "db_session_id": "JRI664OFBPNTT9XIQEES", "orig_file_number": 8, "seqno_to_time_mapping": "N/A"}} 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773088677846018, "job": 1, "event": "recovery_finished"} 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: [db/version_set.cc:5047] Creating manifest 10 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-c/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55614fb48e00 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: DB pointer 0x55614fb58000 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c does not exist in monmap, will attempt to join an existing cluster 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: using public_addr v2:192.168.123.108:0/0 -> [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: ** DB Stats ** 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: ** Compaction Stats [default] ** 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: L0 1/0 1.73 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 3.1 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Sum 1/0 1.73 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 3.1 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 3.1 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: ** Compaction Stats [default] ** 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 3.1 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-09T20:37:58.076 INFO:journalctl@ceph.mon.c.vm08.stdout: Cumulative compaction: 0.00 GB write, 0.35 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout: Interval compaction: 0.00 GB write, 0.35 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout: Block cache BinnedLRUCache@0x55614fb469b0#2 capacity: 512.00 MB usage: 0.98 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 7e-06 secs_since: 0 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout: Block cache entry stats(count,size,portion): DataBlock(1,0.77 KB,0.000146031%) FilterBlock(1,0.11 KB,2.08616e-05%) IndexBlock(1,0.11 KB,2.08616e-05%) Misc(1,0.00 KB,0%) 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout: 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: starting mon.c rank -1 at public addrs [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] at bind addrs [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon_data /var/lib/ceph/mon/ceph-c fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c@-1(???) e0 preinit fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c@-1(synchronizing).mds e1 new map 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c@-1(synchronizing).mds e1 print_map 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout: e1 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout: btime 2026-03-09T20:37:15:402656+0000 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout: legacy client fscid: -1 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout: 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout: No filesystems configured 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c@-1(synchronizing).osd e4 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mkfs 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: monmap epoch 1 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: last_changed 2026-03-09T20:37:14.248038+0000 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: created 2026-03-09T20:37:14.248038+0000 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: min_mon_release 19 (squid) 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: election_strategy: 1 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: fsmap 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: osdmap e1: 0 total, 0 up, 0 in 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mgrmap e1: no daemons active 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/3324809526' entity='client.admin' 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1146515697' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1571332773' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Activating manager daemon a 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mgrmap e2: a(active, starting, since 0.00514513s) 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Manager daemon a is now available 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14100 192.168.123.103:0/1447422968' entity='mgr.a' 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mgrmap e3: a(active, since 1.01238s) 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/3834801571' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/3424911298' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/3424911298' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-09T20:37:58.077 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1290895205' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1290895205' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mgrmap e4: a(active, since 2s) 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1946942812' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Active manager daemon a restarted 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Activating manager daemon a 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: osdmap e2: 0 total, 0 up, 0 in 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mgrmap e5: a(active, starting, since 0.231904s) 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Manager daemon a is now available 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Found migration_current of "None". Setting to last migration. 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mgrmap e6: a(active, since 1.23515s) 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: [09/Mar/2026:20:37:29] ENGINE Bus STARTING 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: [09/Mar/2026:20:37:29] ENGINE Serving on http://192.168.123.103:8765 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: [09/Mar/2026:20:37:29] ENGINE Serving on https://192.168.123.103:7150 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: [09/Mar/2026:20:37:29] ENGINE Bus STARTED 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: [09/Mar/2026:20:37:29] ENGINE Client ('192.168.123.103', 35350) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Generating ssh key... 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mgrmap e7: a(active, since 2s) 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm03", "addr": "192.168.123.103", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Deploying cephadm binary to vm03 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Added host vm03 2026-03-09T20:37:58.078 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Saving service mon spec with placement count:5 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1373298368' entity='client.admin' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/3194674838' entity='client.admin' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Saving service mgr spec with placement count:2 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/352927836' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14118 192.168.123.103:0/4141749778' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]': finished 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/352927836' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mgrmap e8: a(active, since 6s) 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/195646652' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Active manager daemon a restarted 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Activating manager daemon a 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: osdmap e3: 0 total, 0 up, 0 in 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mgrmap e9: a(active, starting, since 0.00536758s) 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Manager daemon a is now available 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mgrmap e10: a(active, since 1.0097s) 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: [09/Mar/2026:20:37:38] ENGINE Bus STARTING 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: [09/Mar/2026:20:37:38] ENGINE Serving on https://192.168.123.103:7150 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: [09/Mar/2026:20:37:38] ENGINE Client ('192.168.123.103', 35366) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: [09/Mar/2026:20:37:38] ENGINE Serving on http://192.168.123.103:8765 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: [09/Mar/2026:20:37:38] ENGINE Bus STARTED 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Deploying daemon agent.vm03 on vm03 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/3271032368' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/3807724702' entity='client.admin' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mgrmap e11: a(active, since 3s) 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.079 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1393567455' entity='client.admin' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm04", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Deploying cephadm binary to vm04 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mgrmap e12: a(active, since 6s) 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Added host vm04 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm04", "caps": []}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm04", "caps": []}]': finished 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm08", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Deploying cephadm binary to vm08 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Deploying daemon agent.vm04 on vm04 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.080 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Added host vm08 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm08", "caps": []}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm08", "caps": []}]': finished 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1248661400' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Deploying daemon agent.vm08 on vm08 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1248661400' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm03:192.168.123.103=a;vm04:192.168.123.104=b;vm08:192.168.123.108=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Saving service mon spec with placement vm03:192.168.123.103=a;vm04:192.168.123.104=b;vm08:192.168.123.108=c;count:3 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Filtered out host vm08: does not belong to mon public_network(s): 192.168.123.0/24, host network(s): 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Deploying daemon mon.b on vm04 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='client.? 192.168.123.108:0/436973558' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.a calling monitor election 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.b calling monitor election 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:58.081 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: monmap epoch 2 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: last_changed 2026-03-09T20:37:50.953583+0000 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: created 2026-03-09T20:37:14.248038+0000 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: min_mon_release 19 (squid) 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: election_strategy: 1 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: 1: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: fsmap 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mgrmap e12: a(active, since 18s) 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: overall HEALTH_OK 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:37:58.082 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:37:57 vm08 ceph-mon[54477]: mon.c@-1(synchronizing).paxosservice(auth 1..6) refresh upgraded, format 0 -> 3 2026-03-09T20:37:58.841 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-09T20:37:58.841 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph mon dump -f json 2026-03-09T20:37:59.096 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.c/config 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: mon.b calling monitor election 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: mon.a calling monitor election 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: monmap epoch 3 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: last_changed 2026-03-09T20:37:57.889003+0000 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: created 2026-03-09T20:37:14.248038+0000 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: min_mon_release 19 (squid) 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: election_strategy: 1 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: 1: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: 2: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: fsmap 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: mgrmap e12: a(active, since 25s) 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: overall HEALTH_OK 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:03.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:03.256 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:03.256 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:03.256 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:02 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: mon.b calling monitor election 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: mon.a calling monitor election 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: monmap epoch 3 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: last_changed 2026-03-09T20:37:57.889003+0000 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: created 2026-03-09T20:37:14.248038+0000 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: min_mon_release 19 (squid) 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: election_strategy: 1 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: 1: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: 2: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: fsmap 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: mgrmap e12: a(active, since 25s) 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: overall HEALTH_OK 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:03.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:02 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:03.790 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:38:03.791 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":3,"fsid":"8176208c-1bf7-11f1-86ba-4f626c793dbd","modified":"2026-03-09T20:37:57.889003Z","created":"2026-03-09T20:37:14.248038Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:3300","nonce":0},{"type":"v1","addr":"192.168.123.103:6789","nonce":0}]},"addr":"192.168.123.103:6789/0","public_addr":"192.168.123.103:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:3300","nonce":0},{"type":"v1","addr":"192.168.123.104:6789","nonce":0}]},"addr":"192.168.123.104:6789/0","public_addr":"192.168.123.104:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":2,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-09T20:38:03.791 INFO:teuthology.orchestra.run.vm08.stderr:dumped monmap epoch 3 2026-03-09T20:38:03.847 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-09T20:38:03.847 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph config generate-minimal-conf 2026-03-09T20:38:04.046 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: Reconfiguring mon.a (unknown last config time)... 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: Reconfiguring daemon mon.a on vm03 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='client.? 192.168.123.108:0/2271932945' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:38:04.126 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:04 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: Reconfiguring mon.a (unknown last config time)... 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: Reconfiguring daemon mon.a on vm03 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='client.? 192.168.123.108:0/2271932945' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:38:04.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:04.256 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.256 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.256 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:38:04.256 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:38:04.256 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:04 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:04.285 INFO:teuthology.orchestra.run.vm03.stdout:# minimal ceph.conf for 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:38:04.285 INFO:teuthology.orchestra.run.vm03.stdout:[global] 2026-03-09T20:38:04.285 INFO:teuthology.orchestra.run.vm03.stdout: fsid = 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:38:04.286 INFO:teuthology.orchestra.run.vm03.stdout: mon_host = [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] 2026-03-09T20:38:04.357 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-09T20:38:04.357 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T20:38:04.358 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T20:38:04.385 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T20:38:04.401 DEBUG:teuthology.orchestra.run.vm03:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:38:04.453 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T20:38:04.453 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T20:38:04.483 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T20:38:04.484 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:38:04.552 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:38:04.552 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T20:38:04.584 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:38:04.584 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:38:04.652 INFO:tasks.cephadm:Adding mgr.a on vm03 2026-03-09T20:38:04.652 INFO:tasks.cephadm:Adding mgr.b on vm04 2026-03-09T20:38:04.652 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph orch apply mgr '2;vm03=a;vm04=b' 2026-03-09T20:38:04.875 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.c/config 2026-03-09T20:38:04.900 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: mon.b calling monitor election 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: mon.a calling monitor election 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: monmap epoch 3 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: last_changed 2026-03-09T20:37:57.889003+0000 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: created 2026-03-09T20:37:14.248038+0000 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: min_mon_release 19 (squid) 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: election_strategy: 1 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: 1: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: 2: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: fsmap 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: mgrmap e12: a(active, since 25s) 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: overall HEALTH_OK 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.901 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: Reconfiguring mon.a (unknown last config time)... 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: Reconfiguring daemon mon.a on vm03 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='client.? 192.168.123.108:0/2271932945' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T20:38:04.902 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:04 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:05.135 INFO:teuthology.orchestra.run.vm08.stdout:Scheduled mgr update... 2026-03-09T20:38:05.212 DEBUG:teuthology.orchestra.run.vm04:mgr.b> sudo journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mgr.b.service 2026-03-09T20:38:05.214 INFO:tasks.cephadm:Deploying OSDs... 2026-03-09T20:38:05.214 DEBUG:teuthology.orchestra.run.vm03:> set -ex 2026-03-09T20:38:05.214 DEBUG:teuthology.orchestra.run.vm03:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T20:38:05.231 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:38:05.231 DEBUG:teuthology.orchestra.run.vm03:> ls /dev/[sv]d? 2026-03-09T20:38:05.290 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vda 2026-03-09T20:38:05.290 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vdb 2026-03-09T20:38:05.290 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vdc 2026-03-09T20:38:05.290 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vdd 2026-03-09T20:38:05.290 INFO:teuthology.orchestra.run.vm03.stdout:/dev/vde 2026-03-09T20:38:05.290 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T20:38:05.290 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T20:38:05.290 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vdb 2026-03-09T20:38:05.351 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vdb 2026-03-09T20:38:05.351 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:38:05.351 INFO:teuthology.orchestra.run.vm03.stdout:Device: 6h/6d Inode: 221 Links: 1 Device type: fc,10 2026-03-09T20:38:05.351 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:38:05.351 INFO:teuthology.orchestra.run.vm03.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:38:05.351 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-09 20:38:02.249234283 +0000 2026-03-09T20:38:05.351 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-09 20:33:54.428630278 +0000 2026-03-09T20:38:05.351 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-09 20:33:54.428630278 +0000 2026-03-09T20:38:05.351 INFO:teuthology.orchestra.run.vm03.stdout: Birth: 2026-03-09 20:31:07.215000000 +0000 2026-03-09T20:38:05.351 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T20:38:05.420 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-09T20:38:05.420 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-09T20:38:05.420 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000170781 s, 3.0 MB/s 2026-03-09T20:38:05.422 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T20:38:05.481 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vdc 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: mon.c calling monitor election 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: mon.c calling monitor election 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: mon.b calling monitor election 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: mon.a calling monitor election 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: mon.a is new leader, mons a,b,c in quorum (ranks 0,1,2) 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: monmap epoch 3 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: last_changed 2026-03-09T20:37:57.889003+0000 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: created 2026-03-09T20:37:14.248038+0000 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: min_mon_release 19 (squid) 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: election_strategy: 1 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: 1: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: 2: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: fsmap 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: mgrmap e12: a(active, since 27s) 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: overall HEALTH_OK 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:38:05.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:05 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:05.540 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vdc 2026-03-09T20:38:05.540 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:38:05.540 INFO:teuthology.orchestra.run.vm03.stdout:Device: 6h/6d Inode: 224 Links: 1 Device type: fc,20 2026-03-09T20:38:05.540 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:38:05.540 INFO:teuthology.orchestra.run.vm03.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:38:05.540 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-09 20:38:02.254234285 +0000 2026-03-09T20:38:05.540 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-09 20:33:54.446630299 +0000 2026-03-09T20:38:05.540 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-09 20:33:54.446630299 +0000 2026-03-09T20:38:05.540 INFO:teuthology.orchestra.run.vm03.stdout: Birth: 2026-03-09 20:31:07.221000000 +0000 2026-03-09T20:38:05.540 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T20:38:05.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: mon.c calling monitor election 2026-03-09T20:38:05.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: mon.c calling monitor election 2026-03-09T20:38:05.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: mon.b calling monitor election 2026-03-09T20:38:05.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: mon.a calling monitor election 2026-03-09T20:38:05.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: mon.a is new leader, mons a,b,c in quorum (ranks 0,1,2) 2026-03-09T20:38:05.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: monmap epoch 3 2026-03-09T20:38:05.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:38:05.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: last_changed 2026-03-09T20:37:57.889003+0000 2026-03-09T20:38:05.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: created 2026-03-09T20:37:14.248038+0000 2026-03-09T20:38:05.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: min_mon_release 19 (squid) 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: election_strategy: 1 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: 1: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: 2: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: fsmap 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: mgrmap e12: a(active, since 27s) 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: overall HEALTH_OK 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:38:05.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:05 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:05.603 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: mon.c calling monitor election 2026-03-09T20:38:05.603 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: mon.c calling monitor election 2026-03-09T20:38:05.603 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: mon.b calling monitor election 2026-03-09T20:38:05.603 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: mon.a calling monitor election 2026-03-09T20:38:05.603 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: mon.a is new leader, mons a,b,c in quorum (ranks 0,1,2) 2026-03-09T20:38:05.603 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: monmap epoch 3 2026-03-09T20:38:05.603 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:38:05.603 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: last_changed 2026-03-09T20:37:57.889003+0000 2026-03-09T20:38:05.603 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: created 2026-03-09T20:37:14.248038+0000 2026-03-09T20:38:05.603 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: min_mon_release 19 (squid) 2026-03-09T20:38:05.603 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: election_strategy: 1 2026-03-09T20:38:05.603 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: 0: [v2:192.168.123.103:3300/0,v1:192.168.123.103:6789/0] mon.a 2026-03-09T20:38:05.603 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: 1: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T20:38:05.603 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: 2: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.c 2026-03-09T20:38:05.604 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: fsmap 2026-03-09T20:38:05.604 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T20:38:05.604 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: mgrmap e12: a(active, since 27s) 2026-03-09T20:38:05.604 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: overall HEALTH_OK 2026-03-09T20:38:05.604 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:05.604 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:05.604 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:05.604 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:05.604 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:05.604 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:38:05.604 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-09T20:38:05.604 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:38:05.604 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:05 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:05.606 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-09T20:38:05.606 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-09T20:38:05.606 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000183273 s, 2.8 MB/s 2026-03-09T20:38:05.607 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T20:38:05.669 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vdd 2026-03-09T20:38:05.732 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vdd 2026-03-09T20:38:05.732 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:38:05.732 INFO:teuthology.orchestra.run.vm03.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,30 2026-03-09T20:38:05.732 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:38:05.732 INFO:teuthology.orchestra.run.vm03.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:38:05.732 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-09 20:38:02.258234286 +0000 2026-03-09T20:38:05.732 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-09 20:33:54.425630275 +0000 2026-03-09T20:38:05.732 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-09 20:33:54.425630275 +0000 2026-03-09T20:38:05.732 INFO:teuthology.orchestra.run.vm03.stdout: Birth: 2026-03-09 20:31:07.234000000 +0000 2026-03-09T20:38:05.732 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T20:38:05.799 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-09T20:38:05.800 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-09T20:38:05.800 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 9.9927e-05 s, 5.1 MB/s 2026-03-09T20:38:05.801 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T20:38:05.862 DEBUG:teuthology.orchestra.run.vm03:> stat /dev/vde 2026-03-09T20:38:05.924 INFO:teuthology.orchestra.run.vm03.stdout: File: /dev/vde 2026-03-09T20:38:05.924 INFO:teuthology.orchestra.run.vm03.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:38:05.924 INFO:teuthology.orchestra.run.vm03.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T20:38:05.924 INFO:teuthology.orchestra.run.vm03.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:38:05.924 INFO:teuthology.orchestra.run.vm03.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:38:05.924 INFO:teuthology.orchestra.run.vm03.stdout:Access: 2026-03-09 20:38:02.262234288 +0000 2026-03-09T20:38:05.924 INFO:teuthology.orchestra.run.vm03.stdout:Modify: 2026-03-09 20:33:54.426630276 +0000 2026-03-09T20:38:05.924 INFO:teuthology.orchestra.run.vm03.stdout:Change: 2026-03-09 20:33:54.426630276 +0000 2026-03-09T20:38:05.924 INFO:teuthology.orchestra.run.vm03.stdout: Birth: 2026-03-09 20:31:07.238000000 +0000 2026-03-09T20:38:05.925 DEBUG:teuthology.orchestra.run.vm03:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T20:38:06.014 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records in 2026-03-09T20:38:06.014 INFO:teuthology.orchestra.run.vm03.stderr:1+0 records out 2026-03-09T20:38:06.014 INFO:teuthology.orchestra.run.vm03.stderr:512 bytes copied, 0.000117489 s, 4.4 MB/s 2026-03-09T20:38:06.015 DEBUG:teuthology.orchestra.run.vm03:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T20:38:06.035 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T20:38:06.035 DEBUG:teuthology.orchestra.run.vm04:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T20:38:06.052 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:05 vm04 systemd[1]: Started Ceph mgr.b for 8176208c-1bf7-11f1-86ba-4f626c793dbd. 2026-03-09T20:38:06.056 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:38:06.056 DEBUG:teuthology.orchestra.run.vm04:> ls /dev/[sv]d? 2026-03-09T20:38:06.116 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vda 2026-03-09T20:38:06.117 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdb 2026-03-09T20:38:06.117 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdc 2026-03-09T20:38:06.117 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdd 2026-03-09T20:38:06.117 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vde 2026-03-09T20:38:06.117 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T20:38:06.117 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T20:38:06.117 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdb 2026-03-09T20:38:06.178 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdb 2026-03-09T20:38:06.178 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:38:06.178 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 253 Links: 1 Device type: fc,10 2026-03-09T20:38:06.178 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:38:06.178 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:38:06.178 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-09 20:37:52.392368485 +0000 2026-03-09T20:38:06.178 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-09 20:33:54.150090075 +0000 2026-03-09T20:38:06.178 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-09 20:33:54.150090075 +0000 2026-03-09T20:38:06.178 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-09 20:30:36.203000000 +0000 2026-03-09T20:38:06.178 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T20:38:06.248 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-09T20:38:06.248 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-09T20:38:06.248 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.00012295 s, 4.2 MB/s 2026-03-09T20:38:06.249 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T20:38:06.311 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdc 2026-03-09T20:38:06.351 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:06.051+0000 7ff4d40eb140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T20:38:06.351 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:06.102+0000 7ff4d40eb140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T20:38:06.376 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdc 2026-03-09T20:38:06.376 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:38:06.376 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-09T20:38:06.376 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:38:06.376 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:38:06.376 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-09 20:37:52.396368485 +0000 2026-03-09T20:38:06.376 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-09 20:33:54.122090035 +0000 2026-03-09T20:38:06.376 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-09 20:33:54.122090035 +0000 2026-03-09T20:38:06.376 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-09 20:30:36.208000000 +0000 2026-03-09T20:38:06.377 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T20:38:06.447 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-09T20:38:06.447 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-09T20:38:06.447 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000184054 s, 2.8 MB/s 2026-03-09T20:38:06.448 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T20:38:06.512 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdd 2026-03-09T20:38:06.572 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdd 2026-03-09T20:38:06.572 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:38:06.572 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T20:38:06.572 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:38:06.572 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:38:06.572 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-09 20:37:52.400368484 +0000 2026-03-09T20:38:06.573 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-09 20:33:54.132090049 +0000 2026-03-09T20:38:06.573 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-09 20:33:54.132090049 +0000 2026-03-09T20:38:06.573 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-09 20:30:36.211000000 +0000 2026-03-09T20:38:06.573 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: from='client.14208 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm03=a;vm04=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: Saving service mgr spec with placement vm03=a;vm04=b;count:2 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: Deploying daemon mgr.b on vm04 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:38:06.611 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: from='client.14208 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm03=a;vm04=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: Saving service mgr spec with placement vm03=a;vm04=b;count:2 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: Deploying daemon mgr.b on vm04 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:38:06.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:06.658 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-09T20:38:06.658 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-09T20:38:06.658 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000199373 s, 2.6 MB/s 2026-03-09T20:38:06.659 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T20:38:06.684 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vde 2026-03-09T20:38:06.764 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vde 2026-03-09T20:38:06.764 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:38:06.764 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T20:38:06.764 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:38:06.764 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:38:06.764 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-09 20:37:52.404368484 +0000 2026-03-09T20:38:06.764 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-09 20:33:54.115090025 +0000 2026-03-09T20:38:06.764 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-09 20:33:54.115090025 +0000 2026-03-09T20:38:06.764 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-09 20:30:36.213000000 +0000 2026-03-09T20:38:06.764 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T20:38:06.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: from='client.14208 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm03=a;vm04=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:38:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: Saving service mgr spec with placement vm03=a;vm04=b;count:2 2026-03-09T20:38:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: Deploying daemon mgr.b on vm04 2026-03-09T20:38:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T20:38:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T20:38:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:06 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:06.863 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-09T20:38:06.863 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-09T20:38:06.863 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.00359956 s, 142 kB/s 2026-03-09T20:38:06.863 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:06 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:06.609+0000 7ff4d40eb140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T20:38:06.864 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T20:38:06.884 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T20:38:06.884 DEBUG:teuthology.orchestra.run.vm08:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T20:38:06.900 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:38:06.900 DEBUG:teuthology.orchestra.run.vm08:> ls /dev/[sv]d? 2026-03-09T20:38:06.959 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vda 2026-03-09T20:38:06.960 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdb 2026-03-09T20:38:06.960 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdc 2026-03-09T20:38:06.960 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdd 2026-03-09T20:38:06.960 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vde 2026-03-09T20:38:06.960 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T20:38:06.960 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T20:38:06.960 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdb 2026-03-09T20:38:07.021 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdb 2026-03-09T20:38:07.021 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:38:07.021 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-09T20:38:07.021 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:38:07.021 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:38:07.021 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-09 20:37:59.405671525 +0000 2026-03-09T20:38:07.021 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-09 20:33:53.946496396 +0000 2026-03-09T20:38:07.021 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-09 20:33:53.946496396 +0000 2026-03-09T20:38:07.021 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-09 20:30:11.243000000 +0000 2026-03-09T20:38:07.021 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T20:38:07.091 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-09T20:38:07.091 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-09T20:38:07.091 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000153627 s, 3.3 MB/s 2026-03-09T20:38:07.093 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T20:38:07.144 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:06 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:06.888+0000 7fd38054f640 -1 mgr.server handle_report got status from non-daemon mon.c 2026-03-09T20:38:07.153 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdc 2026-03-09T20:38:07.215 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdc 2026-03-09T20:38:07.215 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:38:07.215 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-09T20:38:07.215 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:38:07.215 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:38:07.215 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-09 20:37:59.409671529 +0000 2026-03-09T20:38:07.215 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-09 20:33:53.946496396 +0000 2026-03-09T20:38:07.215 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-09 20:33:53.946496396 +0000 2026-03-09T20:38:07.215 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-09 20:30:11.246000000 +0000 2026-03-09T20:38:07.215 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T20:38:07.282 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-09T20:38:07.283 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-09T20:38:07.283 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.00019802 s, 2.6 MB/s 2026-03-09T20:38:07.284 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T20:38:07.344 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdd 2026-03-09T20:38:07.404 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdd 2026-03-09T20:38:07.405 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:38:07.405 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T20:38:07.405 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:38:07.405 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:38:07.405 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-09 20:37:59.413671533 +0000 2026-03-09T20:38:07.405 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-09 20:33:53.968496442 +0000 2026-03-09T20:38:07.405 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-09 20:33:53.968496442 +0000 2026-03-09T20:38:07.405 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-09 20:30:11.249000000 +0000 2026-03-09T20:38:07.405 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-mon[53250]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-mon[53250]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-mon[53250]: Reconfiguring mgr.a (unknown last config time)... 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-mon[53250]: Reconfiguring daemon mgr.a on vm03 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:07.133+0000 7ff4d40eb140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: from numpy import show_config as show_numpy_config 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:07.272+0000 7ff4d40eb140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T20:38:07.459 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:07.321+0000 7ff4d40eb140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T20:38:07.473 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-09T20:38:07.474 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-09T20:38:07.474 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000204833 s, 2.5 MB/s 2026-03-09T20:38:07.475 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T20:38:07.533 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vde 2026-03-09T20:38:07.591 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vde 2026-03-09T20:38:07.591 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T20:38:07.591 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T20:38:07.591 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T20:38:07.591 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T20:38:07.591 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-09 20:37:59.417671536 +0000 2026-03-09T20:38:07.591 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-09 20:33:53.972496450 +0000 2026-03-09T20:38:07.591 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-09 20:33:53.972496450 +0000 2026-03-09T20:38:07.591 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-09 20:30:11.319000000 +0000 2026-03-09T20:38:07.591 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T20:38:07.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:07 vm03 ceph-mon[52226]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:07.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:07 vm03 ceph-mon[52226]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:38:07.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:07 vm03 ceph-mon[52226]: Reconfiguring mgr.a (unknown last config time)... 2026-03-09T20:38:07.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:07 vm03 ceph-mon[52226]: Reconfiguring daemon mgr.a on vm03 2026-03-09T20:38:07.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:07 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:07.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:07 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:07.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:07 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:07.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:07 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:07.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:07 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:07.645 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:07 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:07.657 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:07 vm08 ceph-mon[54477]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:07.658 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:07 vm08 ceph-mon[54477]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:38:07.658 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:07 vm08 ceph-mon[54477]: Reconfiguring mgr.a (unknown last config time)... 2026-03-09T20:38:07.658 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:07 vm08 ceph-mon[54477]: Reconfiguring daemon mgr.a on vm03 2026-03-09T20:38:07.658 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:07 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:07.658 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:07 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:07.658 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:07 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:07.658 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:07 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:07.658 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:07 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:07.658 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:07 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:07.660 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-09T20:38:07.660 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-09T20:38:07.660 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000175048 s, 2.9 MB/s 2026-03-09T20:38:07.661 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T20:38:07.725 INFO:tasks.cephadm:Deploying osd.0 on vm03 with /dev/vde... 2026-03-09T20:38:07.726 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- lvm zap /dev/vde 2026-03-09T20:38:07.744 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:07 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:07.456+0000 7ff4d40eb140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T20:38:07.918 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:08.366 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:08.107+0000 7ff4d40eb140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T20:38:08.366 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:08.220+0000 7ff4d40eb140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:38:08.366 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:08.262+0000 7ff4d40eb140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T20:38:08.366 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:08.303+0000 7ff4d40eb140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T20:38:08.366 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:08.352+0000 7ff4d40eb140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T20:38:08.367 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:08 vm03 ceph-mon[52226]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:38:08.367 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:08 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.367 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:08 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-mon[53250]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-mon[53250]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:08.396+0000 7ff4d40eb140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T20:38:08.630 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:08.571+0000 7ff4d40eb140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T20:38:08.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:08 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:08.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:08 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:08.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:08 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:08 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:08.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:08 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:08 vm03 ceph-mon[52226]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:08.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:08 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:08 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:08 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:08 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.754 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:38:08.778 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph orch daemon add osd vm03:/dev/vde 2026-03-09T20:38:08.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:08 vm08 ceph-mon[54477]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:38:08.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:08 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:08 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:08 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:08.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:08 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:08.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:08 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:08 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:08.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:08 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:08 vm08 ceph-mon[54477]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:08.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:08 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:08 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:08 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:08 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:08.916 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:08.629+0000 7ff4d40eb140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T20:38:08.916 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:08 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:08.868+0000 7ff4d40eb140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T20:38:08.958 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:09.468 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:09 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:09.184+0000 7ff4d40eb140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T20:38:09.468 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:09 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:09.224+0000 7ff4d40eb140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T20:38:09.468 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:09 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:09.271+0000 7ff4d40eb140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T20:38:09.468 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:09 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:09.348+0000 7ff4d40eb140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T20:38:09.468 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:09 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:09.385+0000 7ff4d40eb140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T20:38:09.722 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:09 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:09.466+0000 7ff4d40eb140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T20:38:09.722 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:09 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:09.582+0000 7ff4d40eb140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:38:10.005 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:09 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:09.721+0000 7ff4d40eb140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T20:38:10.005 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:09 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:09.761+0000 7ff4d40eb140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T20:38:10.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:10 vm03 ceph-mon[52226]: from='client.14220 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm03:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:38:10.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:10 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:38:10.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:10 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:38:10.395 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:10 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:10.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:10 vm04 ceph-mon[53250]: from='client.14220 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm03:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:38:10.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:10 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:38:10.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:10 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:38:10.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:10 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:10.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:10 vm08 ceph-mon[54477]: from='client.14220 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm03:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:38:10.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:10 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:38:10.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:10 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:38:10.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:10 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:11.097 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:11 vm03 ceph-mon[52226]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:11.098 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:11 vm03 ceph-mon[52226]: from='mgr.? 192.168.123.104:0/2706667542' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T20:38:11.098 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:11 vm03 ceph-mon[52226]: from='mgr.? 192.168.123.104:0/2706667542' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T20:38:11.098 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:11 vm03 ceph-mon[52226]: from='mgr.? 192.168.123.104:0/2706667542' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T20:38:11.098 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:11 vm03 ceph-mon[52226]: from='mgr.? 192.168.123.104:0/2706667542' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T20:38:11.098 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:11 vm03 ceph-mon[52226]: Standby manager daemon b started 2026-03-09T20:38:11.098 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:11 vm03 ceph-mon[52226]: mgrmap e13: a(active, since 32s), standbys: b 2026-03-09T20:38:11.098 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:11 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T20:38:11.098 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:11 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/2580282234' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "fe57b656-0d2c-4d80-8c77-0f940eeaf7e5"}]: dispatch 2026-03-09T20:38:11.098 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:11 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/2580282234' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "fe57b656-0d2c-4d80-8c77-0f940eeaf7e5"}]': finished 2026-03-09T20:38:11.098 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:11 vm03 ceph-mon[52226]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T20:38:11.098 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:11 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:11.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:11 vm04 ceph-mon[53250]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:11.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:11 vm04 ceph-mon[53250]: from='mgr.? 192.168.123.104:0/2706667542' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T20:38:11.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:11 vm04 ceph-mon[53250]: from='mgr.? 192.168.123.104:0/2706667542' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T20:38:11.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:11 vm04 ceph-mon[53250]: from='mgr.? 192.168.123.104:0/2706667542' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T20:38:11.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:11 vm04 ceph-mon[53250]: from='mgr.? 192.168.123.104:0/2706667542' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T20:38:11.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:11 vm04 ceph-mon[53250]: Standby manager daemon b started 2026-03-09T20:38:11.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:11 vm04 ceph-mon[53250]: mgrmap e13: a(active, since 32s), standbys: b 2026-03-09T20:38:11.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:11 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T20:38:11.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:11 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/2580282234' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "fe57b656-0d2c-4d80-8c77-0f940eeaf7e5"}]: dispatch 2026-03-09T20:38:11.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:11 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/2580282234' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "fe57b656-0d2c-4d80-8c77-0f940eeaf7e5"}]': finished 2026-03-09T20:38:11.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:11 vm04 ceph-mon[53250]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T20:38:11.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:11 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:11.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:11 vm08 ceph-mon[54477]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:11.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:11 vm08 ceph-mon[54477]: from='mgr.? 192.168.123.104:0/2706667542' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T20:38:11.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:11 vm08 ceph-mon[54477]: from='mgr.? 192.168.123.104:0/2706667542' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T20:38:11.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:11 vm08 ceph-mon[54477]: from='mgr.? 192.168.123.104:0/2706667542' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T20:38:11.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:11 vm08 ceph-mon[54477]: from='mgr.? 192.168.123.104:0/2706667542' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T20:38:11.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:11 vm08 ceph-mon[54477]: Standby manager daemon b started 2026-03-09T20:38:11.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:11 vm08 ceph-mon[54477]: mgrmap e13: a(active, since 32s), standbys: b 2026-03-09T20:38:11.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:11 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T20:38:11.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:11 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/2580282234' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "fe57b656-0d2c-4d80-8c77-0f940eeaf7e5"}]: dispatch 2026-03-09T20:38:11.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:11 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/2580282234' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "fe57b656-0d2c-4d80-8c77-0f940eeaf7e5"}]': finished 2026-03-09T20:38:11.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:11 vm08 ceph-mon[54477]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T20:38:11.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:11 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:12.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:12 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/2602286940' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:38:12.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:12 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/2602286940' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:38:12.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:12 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/2602286940' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:38:13.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:13 vm03 ceph-mon[52226]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:13.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:13 vm04 ceph-mon[53250]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:13.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:13 vm08 ceph-mon[54477]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:15.382 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:15 vm03 ceph-mon[52226]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:15.382 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:15 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T20:38:15.382 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:15 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:15.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:15 vm04 ceph-mon[53250]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:15.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:15 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T20:38:15.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:15 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:15.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:15 vm08 ceph-mon[54477]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:15.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:15 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T20:38:15.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:15 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:16.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:16 vm03 ceph-mon[52226]: Deploying daemon osd.0 on vm03 2026-03-09T20:38:16.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:16 vm04 ceph-mon[53250]: Deploying daemon osd.0 on vm03 2026-03-09T20:38:16.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:16 vm08 ceph-mon[54477]: Deploying daemon osd.0 on vm03 2026-03-09T20:38:17.158 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:17 vm03 ceph-mon[52226]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:17.158 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:17 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:17.158 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:17 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:17.158 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:17 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:17.158 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:17 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:17.408 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:17 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:17.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:17 vm04 ceph-mon[53250]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:17.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:17 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:17.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:17 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:17.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:17 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:17.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:17 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:17.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:17 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:17.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:17 vm08 ceph-mon[54477]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:17.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:17 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:17.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:17 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:17.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:17 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:17.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:17 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:17.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:17 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:18.142 INFO:teuthology.orchestra.run.vm03.stdout:Created osd(s) 0 on host 'vm03' 2026-03-09T20:38:18.207 DEBUG:teuthology.orchestra.run.vm03:osd.0> sudo journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@osd.0.service 2026-03-09T20:38:18.210 INFO:tasks.cephadm:Deploying osd.1 on vm04 with /dev/vde... 2026-03-09T20:38:18.211 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- lvm zap /dev/vde 2026-03-09T20:38:18.246 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:18 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:18.246 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:18 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:18.246 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:18 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:18.246 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:18 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:18.246 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:18 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:18.246 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:18 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:18.246 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:18 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:18.246 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:18 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:18.402 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.b/config 2026-03-09T20:38:18.428 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:18 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:18.428 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:18 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:18.428 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:18 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:18.428 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:18 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:18.428 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:18 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:18.428 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:18 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:18.428 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:18 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:18.428 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:18 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:18.491 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:18 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:18.492 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:18 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:18.492 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:18 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:18.492 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:18 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:18.492 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:18 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:18.492 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:18 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:18.492 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:18 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:18.492 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:18 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:19.144 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:38:18 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-0[60583]: 2026-03-09T20:38:18.803+0000 7f3d6d48a740 -1 osd.0 0 log_to_monitors true 2026-03-09T20:38:19.222 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:38:19.246 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph orch daemon add osd vm04:/dev/vde 2026-03-09T20:38:19.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:19 vm08 ceph-mon[54477]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:19.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:19 vm08 ceph-mon[54477]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T20:38:19.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:19 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:19.436 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.b/config 2026-03-09T20:38:19.460 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:19 vm04 ceph-mon[53250]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:19.460 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:19 vm04 ceph-mon[53250]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T20:38:19.460 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:19 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:19.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:19 vm03 ceph-mon[52226]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:19.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:19 vm03 ceph-mon[52226]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T20:38:19.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:19 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:20.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:20 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:20.204 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:20 vm04 ceph-mon[53250]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T20:38:20.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:20 vm04 ceph-mon[53250]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T20:38:20.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:20 vm04 ceph-mon[53250]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-09T20:38:20.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:20 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:20.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:20 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:20.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:20 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:38:20.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:20 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:38:20.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:20 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:20.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:20 vm04 ceph-mon[53250]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-09T20:38:20.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:20 vm04 ceph-mon[53250]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T20:38:20.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:20 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:20.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:20 vm08 ceph-mon[54477]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T20:38:20.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:20 vm08 ceph-mon[54477]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T20:38:20.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:20 vm08 ceph-mon[54477]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-09T20:38:20.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:20 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:20.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:20 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:20.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:20 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:38:20.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:20 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:38:20.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:20 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:20.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:20 vm08 ceph-mon[54477]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-09T20:38:20.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:20 vm08 ceph-mon[54477]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T20:38:20.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:20 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:20.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:20 vm03 ceph-mon[52226]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T20:38:20.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:20 vm03 ceph-mon[52226]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T20:38:20.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:20 vm03 ceph-mon[52226]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm03", "root=default"]}]: dispatch 2026-03-09T20:38:20.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:20 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:20.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:20 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:20.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:20 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:38:20.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:20 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:38:20.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:20 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:20.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:20 vm03 ceph-mon[52226]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm03", "root=default"]}]': finished 2026-03-09T20:38:20.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:20 vm03 ceph-mon[52226]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T20:38:21.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:21 vm04 ceph-mon[53250]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:21.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:21 vm04 ceph-mon[53250]: from='client.24124 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:38:21.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:21 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:21.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:21 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:21.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:21 vm04 ceph-mon[53250]: from='client.? 192.168.123.104:0/4188937771' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d9693d38-ddb7-40f9-bcab-2b2f12cb4552"}]: dispatch 2026-03-09T20:38:21.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:21 vm04 ceph-mon[53250]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d9693d38-ddb7-40f9-bcab-2b2f12cb4552"}]: dispatch 2026-03-09T20:38:21.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:21 vm04 ceph-mon[53250]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "d9693d38-ddb7-40f9-bcab-2b2f12cb4552"}]': finished 2026-03-09T20:38:21.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:21 vm04 ceph-mon[53250]: osdmap e8: 2 total, 0 up, 2 in 2026-03-09T20:38:21.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:21 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:21.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:21 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:21.256 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:21 vm04 ceph-mon[53250]: from='client.? 192.168.123.104:0/221917237' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:38:21.256 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:21 vm04 ceph-mon[53250]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' 2026-03-09T20:38:21.256 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:21 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:21.341 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:38:21 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-0[60583]: 2026-03-09T20:38:21.054+0000 7f3d6940b640 -1 osd.0 0 waiting for initial osdmap 2026-03-09T20:38:21.341 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:38:21 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-0[60583]: 2026-03-09T20:38:21.064+0000 7f3d64a34640 -1 osd.0 8 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T20:38:21.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:21 vm03 ceph-mon[52226]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:21.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:21 vm03 ceph-mon[52226]: from='client.24124 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:38:21.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:21 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:21.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:21 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:21.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:21 vm03 ceph-mon[52226]: from='client.? 192.168.123.104:0/4188937771' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d9693d38-ddb7-40f9-bcab-2b2f12cb4552"}]: dispatch 2026-03-09T20:38:21.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:21 vm03 ceph-mon[52226]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d9693d38-ddb7-40f9-bcab-2b2f12cb4552"}]: dispatch 2026-03-09T20:38:21.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:21 vm03 ceph-mon[52226]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "d9693d38-ddb7-40f9-bcab-2b2f12cb4552"}]': finished 2026-03-09T20:38:21.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:21 vm03 ceph-mon[52226]: osdmap e8: 2 total, 0 up, 2 in 2026-03-09T20:38:21.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:21 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:21.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:21 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:21.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:21 vm03 ceph-mon[52226]: from='client.? 192.168.123.104:0/221917237' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:38:21.342 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:21 vm03 ceph-mon[52226]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' 2026-03-09T20:38:21.342 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:21 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:21.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:21 vm08 ceph-mon[54477]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:21.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:21 vm08 ceph-mon[54477]: from='client.24124 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:38:21.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:21 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:21.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:21 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:21.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:21 vm08 ceph-mon[54477]: from='client.? 192.168.123.104:0/4188937771' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d9693d38-ddb7-40f9-bcab-2b2f12cb4552"}]: dispatch 2026-03-09T20:38:21.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:21 vm08 ceph-mon[54477]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "d9693d38-ddb7-40f9-bcab-2b2f12cb4552"}]: dispatch 2026-03-09T20:38:21.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:21 vm08 ceph-mon[54477]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "d9693d38-ddb7-40f9-bcab-2b2f12cb4552"}]': finished 2026-03-09T20:38:21.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:21 vm08 ceph-mon[54477]: osdmap e8: 2 total, 0 up, 2 in 2026-03-09T20:38:21.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:21 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:21.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:21 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:21.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:21 vm08 ceph-mon[54477]: from='client.? 192.168.123.104:0/221917237' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:38:21.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:21 vm08 ceph-mon[54477]: from='osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582]' entity='osd.0' 2026-03-09T20:38:21.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:21 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:22.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:22 vm03 ceph-mon[52226]: purged_snaps scrub starts 2026-03-09T20:38:22.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:22 vm03 ceph-mon[52226]: purged_snaps scrub ok 2026-03-09T20:38:22.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:22 vm03 ceph-mon[52226]: osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582] boot 2026-03-09T20:38:22.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:22 vm03 ceph-mon[52226]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T20:38:22.341 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:22 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:22.342 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:22 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:22.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:22 vm04 ceph-mon[53250]: purged_snaps scrub starts 2026-03-09T20:38:22.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:22 vm04 ceph-mon[53250]: purged_snaps scrub ok 2026-03-09T20:38:22.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:22 vm04 ceph-mon[53250]: osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582] boot 2026-03-09T20:38:22.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:22 vm04 ceph-mon[53250]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T20:38:22.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:22 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:22.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:22 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:22.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:22 vm08 ceph-mon[54477]: purged_snaps scrub starts 2026-03-09T20:38:22.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:22 vm08 ceph-mon[54477]: purged_snaps scrub ok 2026-03-09T20:38:22.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:22 vm08 ceph-mon[54477]: osd.0 [v2:192.168.123.103:6802/351351582,v1:192.168.123.103:6803/351351582] boot 2026-03-09T20:38:22.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:22 vm08 ceph-mon[54477]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T20:38:22.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:22 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:22.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:22 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:23.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:23 vm03 ceph-mon[52226]: pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:23.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:23 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:23.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:23 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:23.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:23 vm03 ceph-mon[52226]: Detected new or changed devices on vm03 2026-03-09T20:38:23.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:23 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:23.740 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:23 vm04 ceph-mon[53250]: pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:23.741 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:23 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:23.741 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:23 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:23.741 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:23 vm04 ceph-mon[53250]: Detected new or changed devices on vm03 2026-03-09T20:38:23.741 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:23 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:23.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:23 vm08 ceph-mon[54477]: pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T20:38:23.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:23 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:23.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:23 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:23.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:23 vm08 ceph-mon[54477]: Detected new or changed devices on vm03 2026-03-09T20:38:23.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:23 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:24.350 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:24 vm04 ceph-mon[53250]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T20:38:24.350 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:24 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:24.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:24 vm03 ceph-mon[52226]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T20:38:24.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:24 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:24.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:24 vm08 ceph-mon[54477]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T20:38:24.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:24 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:25.515 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:25 vm04 ceph-mon[53250]: pgmap v23: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:25.515 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:25 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T20:38:25.515 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:25 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:25.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:25 vm03 ceph-mon[52226]: pgmap v23: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:25.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:25 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T20:38:25.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:25 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:25.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:25 vm08 ceph-mon[54477]: pgmap v23: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:25.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:25 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T20:38:25.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:25 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:26.616 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:26 vm04 ceph-mon[53250]: Deploying daemon osd.1 on vm04 2026-03-09T20:38:26.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:26 vm03 ceph-mon[52226]: Deploying daemon osd.1 on vm04 2026-03-09T20:38:26.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:26 vm08 ceph-mon[54477]: Deploying daemon osd.1 on vm04 2026-03-09T20:38:27.462 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:27 vm04 ceph-mon[53250]: pgmap v24: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:27.462 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:27 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:27.462 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:27 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:27.462 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:27 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:27.463 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:27 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:27.463 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:27 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:27.463 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:27 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:27.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:27 vm03 ceph-mon[52226]: pgmap v24: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:27.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:27 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:27.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:27 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:27.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:27 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:27.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:27 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:27.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:27 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:27.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:27 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:27.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:27 vm08 ceph-mon[54477]: pgmap v24: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:27.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:27 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:27.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:27 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:27.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:27 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:27.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:27 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:27.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:27 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:27.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:27 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:29.067 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:28 vm08 ceph-mon[54477]: pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:29.069 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:28 vm04 ceph-mon[53250]: pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:29.085 INFO:teuthology.orchestra.run.vm04.stdout:Created osd(s) 1 on host 'vm04' 2026-03-09T20:38:29.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:28 vm03 ceph-mon[52226]: pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:29.159 DEBUG:teuthology.orchestra.run.vm04:osd.1> sudo journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@osd.1.service 2026-03-09T20:38:29.162 INFO:tasks.cephadm:Deploying osd.2 on vm08 with /dev/vde... 2026-03-09T20:38:29.162 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- lvm zap /dev/vde 2026-03-09T20:38:29.366 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.c/config 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: Detected new or changed devices on vm04 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='osd.1 [v2:192.168.123.104:6800/3429737005,v1:192.168.123.104:6801/3429737005]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:29.790 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:29 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: Detected new or changed devices on vm04 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='osd.1 [v2:192.168.123.104:6800/3429737005,v1:192.168.123.104:6801/3429737005]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:30.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:29 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: Detected new or changed devices on vm04 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='osd.1 [v2:192.168.123.104:6800/3429737005,v1:192.168.123.104:6801/3429737005]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:30.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:29 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:30.193 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:38:30.214 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph orch daemon add osd vm08:/dev/vde 2026-03-09T20:38:30.391 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.c/config 2026-03-09T20:38:31.234 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:31 vm08 ceph-mon[54477]: pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:31.234 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:31 vm08 ceph-mon[54477]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T20:38:31.234 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:31 vm08 ceph-mon[54477]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T20:38:31.234 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:31 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:31.234 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:31 vm08 ceph-mon[54477]: from='osd.1 [v2:192.168.123.104:6800/3429737005,v1:192.168.123.104:6801/3429737005]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T20:38:31.234 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:31 vm08 ceph-mon[54477]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T20:38:31.234 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:31 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:38:31.234 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:31 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:38:31.234 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:31 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:31 vm03 ceph-mon[52226]: pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:31 vm03 ceph-mon[52226]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T20:38:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:31 vm03 ceph-mon[52226]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T20:38:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:31 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:31 vm03 ceph-mon[52226]: from='osd.1 [v2:192.168.123.104:6800/3429737005,v1:192.168.123.104:6801/3429737005]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T20:38:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:31 vm03 ceph-mon[52226]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T20:38:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:31 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:38:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:31 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:38:31.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:31 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:31.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:31 vm04 ceph-mon[53250]: pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:31.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:31 vm04 ceph-mon[53250]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T20:38:31.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:31 vm04 ceph-mon[53250]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T20:38:31.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:31 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:31.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:31 vm04 ceph-mon[53250]: from='osd.1 [v2:192.168.123.104:6800/3429737005,v1:192.168.123.104:6801/3429737005]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T20:38:31.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:31 vm04 ceph-mon[53250]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T20:38:31.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:31 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T20:38:31.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:31 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T20:38:31.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:31 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:32.255 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:38:31 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-1[57533]: 2026-03-09T20:38:31.907+0000 7fc0333d0640 -1 osd.1 0 waiting for initial osdmap 2026-03-09T20:38:32.255 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:38:31 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-1[57533]: 2026-03-09T20:38:31.915+0000 7fc02e1e6640 -1 osd.1 13 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T20:38:32.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:32 vm04 ceph-mon[53250]: purged_snaps scrub starts 2026-03-09T20:38:32.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:32 vm04 ceph-mon[53250]: purged_snaps scrub ok 2026-03-09T20:38:32.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:32 vm04 ceph-mon[53250]: from='client.24149 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:38:32.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:32 vm04 ceph-mon[53250]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T20:38:32.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:32 vm04 ceph-mon[53250]: osdmap e12: 2 total, 1 up, 2 in 2026-03-09T20:38:32.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:32 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:32.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:32 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:32.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:32 vm04 ceph-mon[53250]: from='client.? 192.168.123.108:0/3950886816' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a7e4dc58-4f9e-4949-84ac-fa3c14484ade"}]: dispatch 2026-03-09T20:38:32.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:32 vm04 ceph-mon[53250]: from='client.? 192.168.123.108:0/3950886816' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a7e4dc58-4f9e-4949-84ac-fa3c14484ade"}]': finished 2026-03-09T20:38:32.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:32 vm04 ceph-mon[53250]: osdmap e13: 3 total, 1 up, 3 in 2026-03-09T20:38:32.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:32 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:32.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:32 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:32.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:32 vm04 ceph-mon[53250]: from='osd.1 ' entity='osd.1' 2026-03-09T20:38:32.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:32 vm04 ceph-mon[53250]: from='client.? 192.168.123.108:0/3127316857' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:38:32.317 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:32 vm08 ceph-mon[54477]: purged_snaps scrub starts 2026-03-09T20:38:32.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:32 vm08 ceph-mon[54477]: purged_snaps scrub ok 2026-03-09T20:38:32.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:32 vm08 ceph-mon[54477]: from='client.24149 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:38:32.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:32 vm08 ceph-mon[54477]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T20:38:32.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:32 vm08 ceph-mon[54477]: osdmap e12: 2 total, 1 up, 2 in 2026-03-09T20:38:32.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:32 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:32.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:32 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:32.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:32 vm08 ceph-mon[54477]: from='client.? 192.168.123.108:0/3950886816' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a7e4dc58-4f9e-4949-84ac-fa3c14484ade"}]: dispatch 2026-03-09T20:38:32.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:32 vm08 ceph-mon[54477]: from='client.? 192.168.123.108:0/3950886816' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a7e4dc58-4f9e-4949-84ac-fa3c14484ade"}]': finished 2026-03-09T20:38:32.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:32 vm08 ceph-mon[54477]: osdmap e13: 3 total, 1 up, 3 in 2026-03-09T20:38:32.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:32 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:32.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:32 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:32.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:32 vm08 ceph-mon[54477]: from='osd.1 ' entity='osd.1' 2026-03-09T20:38:32.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:32 vm08 ceph-mon[54477]: from='client.? 192.168.123.108:0/3127316857' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:38:32.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:32 vm03 ceph-mon[52226]: purged_snaps scrub starts 2026-03-09T20:38:32.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:32 vm03 ceph-mon[52226]: purged_snaps scrub ok 2026-03-09T20:38:32.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:32 vm03 ceph-mon[52226]: from='client.24149 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:38:32.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:32 vm03 ceph-mon[52226]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T20:38:32.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:32 vm03 ceph-mon[52226]: osdmap e12: 2 total, 1 up, 2 in 2026-03-09T20:38:32.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:32 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:32.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:32 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:32.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:32 vm03 ceph-mon[52226]: from='client.? 192.168.123.108:0/3950886816' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a7e4dc58-4f9e-4949-84ac-fa3c14484ade"}]: dispatch 2026-03-09T20:38:32.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:32 vm03 ceph-mon[52226]: from='client.? 192.168.123.108:0/3950886816' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a7e4dc58-4f9e-4949-84ac-fa3c14484ade"}]': finished 2026-03-09T20:38:32.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:32 vm03 ceph-mon[52226]: osdmap e13: 3 total, 1 up, 3 in 2026-03-09T20:38:32.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:32 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:32.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:32 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:32.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:32 vm03 ceph-mon[52226]: from='osd.1 ' entity='osd.1' 2026-03-09T20:38:32.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:32 vm03 ceph-mon[52226]: from='client.? 192.168.123.108:0/3127316857' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T20:38:33.317 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:33 vm08 ceph-mon[54477]: pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:33.317 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:33 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:33.317 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:33 vm08 ceph-mon[54477]: osd.1 [v2:192.168.123.104:6800/3429737005,v1:192.168.123.104:6801/3429737005] boot 2026-03-09T20:38:33.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:33 vm08 ceph-mon[54477]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T20:38:33.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:33 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:33.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:33 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:33.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:33 vm03 ceph-mon[52226]: pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:33.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:33 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:33.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:33 vm03 ceph-mon[52226]: osd.1 [v2:192.168.123.104:6800/3429737005,v1:192.168.123.104:6801/3429737005] boot 2026-03-09T20:38:33.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:33 vm03 ceph-mon[52226]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T20:38:33.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:33 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:33.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:33 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:33.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:33 vm04 ceph-mon[53250]: pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T20:38:33.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:33 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:33.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:33 vm04 ceph-mon[53250]: osd.1 [v2:192.168.123.104:6800/3429737005,v1:192.168.123.104:6801/3429737005] boot 2026-03-09T20:38:33.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:33 vm04 ceph-mon[53250]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T20:38:33.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:33 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:33.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:33 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:35.216 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:35 vm08 ceph-mon[54477]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:35.216 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:35 vm08 ceph-mon[54477]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T20:38:35.216 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:35 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:35.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:35 vm04 ceph-mon[53250]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:35.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:35 vm04 ceph-mon[53250]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T20:38:35.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:35 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:35.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:35 vm03 ceph-mon[52226]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:35.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:35 vm03 ceph-mon[52226]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T20:38:35.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:35 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:37.316 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:37 vm08 ceph-mon[54477]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:37.316 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:37 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T20:38:37.316 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:37 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:37.316 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:37 vm08 ceph-mon[54477]: Deploying daemon osd.2 on vm08 2026-03-09T20:38:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:37 vm04 ceph-mon[53250]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:37 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T20:38:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:37 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:37 vm04 ceph-mon[53250]: Deploying daemon osd.2 on vm08 2026-03-09T20:38:37.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:37 vm03 ceph-mon[52226]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:37.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:37 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T20:38:37.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:37 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:37.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:37 vm03 ceph-mon[52226]: Deploying daemon osd.2 on vm08 2026-03-09T20:38:38.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:38 vm03 ceph-mon[52226]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:39.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:38 vm04 ceph-mon[53250]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:39.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:38 vm08 ceph-mon[54477]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:39.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:39 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:39.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:39 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:39.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:39 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:40.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:39 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:40.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:39 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:40.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:39 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:40.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:39 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:40.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:39 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:40.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:39 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: pgmap v36: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: pgmap v38: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: pgmap v39: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:41 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: pgmap v36: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: pgmap v38: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: pgmap v39: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.411 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:41 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.493 INFO:teuthology.orchestra.run.vm08.stdout:Created osd(s) 2 on host 'vm08' 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: pgmap v36: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: pgmap v38: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: pgmap v39: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:41.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:41 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:41.571 DEBUG:teuthology.orchestra.run.vm08:osd.2> sudo journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@osd.2.service 2026-03-09T20:38:41.573 INFO:tasks.cephadm:Waiting for 3 OSDs to come up... 2026-03-09T20:38:41.573 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph osd stat -f json 2026-03-09T20:38:41.766 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:41.818 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:38:41 vm08 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-2[58473]: 2026-03-09T20:38:41.639+0000 7f419daf0740 -1 osd.2 0 log_to_monitors true 2026-03-09T20:38:42.009 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:38:42.062 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":15,"num_osds":3,"num_up_osds":2,"osd_up_since":1773088712,"num_in_osds":3,"osd_in_since":1773088711,"num_remapped_pgs":0} 2026-03-09T20:38:42.321 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:42.321 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:42.321 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:42.321 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:42.321 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:42.321 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:42.321 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:42.321 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:42 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:42.321 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:42 vm03 ceph-mon[52226]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T20:38:42.321 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:42 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/117256362' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:38:42.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:42 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:42.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:42 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:42.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:42 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:42.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:42 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:42.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:42 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:42.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:42 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:42.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:42 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:42.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:42 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:42.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:42 vm04 ceph-mon[53250]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T20:38:42.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:42 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/117256362' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:38:42.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:42 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:42.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:42 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:42.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:42 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:42.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:42 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:42.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:42 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:42.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:42 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:42.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:42 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:42.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:42 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:42.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:42 vm08 ceph-mon[54477]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T20:38:42.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:42 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/117256362' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:38:43.063 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph osd stat -f json 2026-03-09T20:38:43.261 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:43.340 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:43 vm03 ceph-mon[52226]: Adjusting osd_memory_target on vm08 to 3329M 2026-03-09T20:38:43.340 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:43 vm03 ceph-mon[52226]: pgmap v41: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:43.340 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:43 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.340 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:43 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.340 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:43 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.340 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:43 vm03 ceph-mon[52226]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T20:38:43.340 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:43 vm03 ceph-mon[52226]: osdmap e16: 3 total, 2 up, 3 in 2026-03-09T20:38:43.340 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:43 vm03 ceph-mon[52226]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T20:38:43.340 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:43 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:43.340 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:43 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.340 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:43 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.340 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:43 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:43 vm04 ceph-mon[53250]: Adjusting osd_memory_target on vm08 to 3329M 2026-03-09T20:38:43.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:43 vm04 ceph-mon[53250]: pgmap v41: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:43.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:43 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:43 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:43 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:43 vm04 ceph-mon[53250]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T20:38:43.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:43 vm04 ceph-mon[53250]: osdmap e16: 3 total, 2 up, 3 in 2026-03-09T20:38:43.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:43 vm04 ceph-mon[53250]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T20:38:43.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:43 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:43.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:43 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:43 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:43 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.520 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:38:43.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:43 vm08 ceph-mon[54477]: Adjusting osd_memory_target on vm08 to 3329M 2026-03-09T20:38:43.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:43 vm08 ceph-mon[54477]: pgmap v41: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:43.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:43 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:43 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:43 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:43 vm08 ceph-mon[54477]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T20:38:43.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:43 vm08 ceph-mon[54477]: osdmap e16: 3 total, 2 up, 3 in 2026-03-09T20:38:43.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:43 vm08 ceph-mon[54477]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T20:38:43.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:43 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:43.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:43 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:43 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:43 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:43.587 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":17,"num_osds":3,"num_up_osds":2,"osd_up_since":1773088712,"num_in_osds":3,"osd_in_since":1773088711,"num_remapped_pgs":0} 2026-03-09T20:38:44.588 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph osd stat -f json 2026-03-09T20:38:44.777 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:44.802 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:44 vm03 ceph-mon[52226]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-09T20:38:44.802 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:44 vm03 ceph-mon[52226]: osdmap e17: 3 total, 2 up, 3 in 2026-03-09T20:38:44.802 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:44 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:44.802 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:44 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/3414655092' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:38:44.802 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:44 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:44.802 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:44 vm03 ceph-mon[52226]: pgmap v44: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:44.802 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:44 vm03 ceph-mon[52226]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' 2026-03-09T20:38:44.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:44 vm08 ceph-mon[54477]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-09T20:38:44.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:44 vm08 ceph-mon[54477]: osdmap e17: 3 total, 2 up, 3 in 2026-03-09T20:38:44.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:44 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:44.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:44 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/3414655092' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:38:44.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:44 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:44.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:44 vm08 ceph-mon[54477]: pgmap v44: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:44.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:44 vm08 ceph-mon[54477]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' 2026-03-09T20:38:44.818 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:38:44 vm08 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-2[58473]: 2026-03-09T20:38:44.392+0000 7f4199a71640 -1 osd.2 0 waiting for initial osdmap 2026-03-09T20:38:44.818 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:38:44 vm08 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-2[58473]: 2026-03-09T20:38:44.401+0000 7f419509a640 -1 osd.2 17 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T20:38:45.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:44 vm04 ceph-mon[53250]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-09T20:38:45.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:44 vm04 ceph-mon[53250]: osdmap e17: 3 total, 2 up, 3 in 2026-03-09T20:38:45.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:44 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:45.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:44 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/3414655092' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:38:45.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:44 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:45.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:44 vm04 ceph-mon[53250]: pgmap v44: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T20:38:45.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:44 vm04 ceph-mon[53250]: from='osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367]' entity='osd.2' 2026-03-09T20:38:45.007 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:38:45.056 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":17,"num_osds":3,"num_up_osds":2,"osd_up_since":1773088712,"num_in_osds":3,"osd_in_since":1773088711,"num_remapped_pgs":0} 2026-03-09T20:38:45.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:45 vm08 ceph-mon[54477]: purged_snaps scrub starts 2026-03-09T20:38:45.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:45 vm08 ceph-mon[54477]: purged_snaps scrub ok 2026-03-09T20:38:45.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:45 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:45.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:45 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/325696630' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:38:45.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:45 vm08 ceph-mon[54477]: osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367] boot 2026-03-09T20:38:45.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:45 vm08 ceph-mon[54477]: osdmap e18: 3 total, 3 up, 3 in 2026-03-09T20:38:45.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:45 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:45.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:45 vm03 ceph-mon[52226]: purged_snaps scrub starts 2026-03-09T20:38:45.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:45 vm03 ceph-mon[52226]: purged_snaps scrub ok 2026-03-09T20:38:45.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:45 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:45.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:45 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/325696630' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:38:45.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:45 vm03 ceph-mon[52226]: osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367] boot 2026-03-09T20:38:45.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:45 vm03 ceph-mon[52226]: osdmap e18: 3 total, 3 up, 3 in 2026-03-09T20:38:45.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:45 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:46.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:45 vm04 ceph-mon[53250]: purged_snaps scrub starts 2026-03-09T20:38:46.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:45 vm04 ceph-mon[53250]: purged_snaps scrub ok 2026-03-09T20:38:46.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:45 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:46.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:45 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/325696630' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:38:46.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:45 vm04 ceph-mon[53250]: osd.2 [v2:192.168.123.108:6800/4157019367,v1:192.168.123.108:6801/4157019367] boot 2026-03-09T20:38:46.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:45 vm04 ceph-mon[53250]: osdmap e18: 3 total, 3 up, 3 in 2026-03-09T20:38:46.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:45 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:46.058 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph osd stat -f json 2026-03-09T20:38:46.244 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:46.483 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:38:46.562 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":18,"num_osds":3,"num_up_osds":3,"osd_up_since":1773088725,"num_in_osds":3,"osd_in_since":1773088711,"num_remapped_pgs":0} 2026-03-09T20:38:46.562 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph osd dump --format=json 2026-03-09T20:38:46.645 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:46 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T20:38:46.645 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:46 vm03 ceph-mon[52226]: pgmap v46: 0 pgs: ; 0 B data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:38:46.645 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:46 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1212958645' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:38:46.755 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:46.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:46 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T20:38:46.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:46 vm08 ceph-mon[54477]: pgmap v46: 0 pgs: ; 0 B data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:38:46.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:46 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1212958645' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:38:46.818 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:46 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T20:38:46.818 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:46 vm04 ceph-mon[53250]: pgmap v46: 0 pgs: ; 0 B data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:38:46.818 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:46 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1212958645' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T20:38:46.990 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:38:46.991 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":19,"fsid":"8176208c-1bf7-11f1-86ba-4f626c793dbd","created":"2026-03-09T20:37:15.404226+0000","modified":"2026-03-09T20:38:46.520776+0000","last_up_change":"2026-03-09T20:38:45.396378+0000","last_in_change":"2026-03-09T20:38:31.419551+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T20:38:45.717276+0000","flags":32769,"flags_names":"hashpspool,creating","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"19","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"fe57b656-0d2c-4d80-8c77-0f940eeaf7e5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6802","nonce":351351582},{"type":"v1","addr":"192.168.123.103:6803","nonce":351351582}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6804","nonce":351351582},{"type":"v1","addr":"192.168.123.103:6805","nonce":351351582}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6808","nonce":351351582},{"type":"v1","addr":"192.168.123.103:6809","nonce":351351582}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6806","nonce":351351582},{"type":"v1","addr":"192.168.123.103:6807","nonce":351351582}]},"public_addr":"192.168.123.103:6803/351351582","cluster_addr":"192.168.123.103:6805/351351582","heartbeat_back_addr":"192.168.123.103:6809/351351582","heartbeat_front_addr":"192.168.123.103:6807/351351582","state":["exists","up"]},{"osd":1,"uuid":"d9693d38-ddb7-40f9-bcab-2b2f12cb4552","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":14,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":3429737005},{"type":"v1","addr":"192.168.123.104:6801","nonce":3429737005}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":3429737005},{"type":"v1","addr":"192.168.123.104:6803","nonce":3429737005}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":3429737005},{"type":"v1","addr":"192.168.123.104:6807","nonce":3429737005}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":3429737005},{"type":"v1","addr":"192.168.123.104:6805","nonce":3429737005}]},"public_addr":"192.168.123.104:6801/3429737005","cluster_addr":"192.168.123.104:6803/3429737005","heartbeat_back_addr":"192.168.123.104:6807/3429737005","heartbeat_front_addr":"192.168.123.104:6805/3429737005","state":["exists","up"]},{"osd":2,"uuid":"a7e4dc58-4f9e-4949-84ac-fa3c14484ade","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":4157019367},{"type":"v1","addr":"192.168.123.108:6801","nonce":4157019367}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":4157019367},{"type":"v1","addr":"192.168.123.108:6803","nonce":4157019367}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":4157019367},{"type":"v1","addr":"192.168.123.108:6807","nonce":4157019367}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":4157019367},{"type":"v1","addr":"192.168.123.108:6805","nonce":4157019367}]},"public_addr":"192.168.123.108:6801/4157019367","cluster_addr":"192.168.123.108:6803/4157019367","heartbeat_back_addr":"192.168.123.108:6807/4157019367","heartbeat_front_addr":"192.168.123.108:6805/4157019367","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:38:19.762215+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:38:29.781918+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:38:42.680940+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.103:6801/512166571":"2026-03-10T20:37:37.656342+0000","192.168.123.103:0/2604231392":"2026-03-10T20:37:37.656342+0000","192.168.123.103:6800/512166571":"2026-03-10T20:37:37.656342+0000","192.168.123.103:0/3477391265":"2026-03-10T20:37:37.656342+0000","192.168.123.103:0/353200345":"2026-03-10T20:37:37.656342+0000","192.168.123.103:0/362558974":"2026-03-10T20:37:27.651886+0000","192.168.123.103:0/2908564848":"2026-03-10T20:37:27.651886+0000","192.168.123.103:0/3602358308":"2026-03-10T20:37:27.651886+0000","192.168.123.103:6801/3012786188":"2026-03-10T20:37:27.651886+0000","192.168.123.103:6800/3012786188":"2026-03-10T20:37:27.651886+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T20:38:47.046 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-09T20:38:45.717276+0000', 'flags': 32769, 'flags_names': 'hashpspool,creating', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '19', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 3, 'score_stable': 3, 'optimal_score': 1, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-09T20:38:47.047 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph osd pool get .mgr pg_num 2026-03-09T20:38:47.230 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:47.470 INFO:teuthology.orchestra.run.vm03.stdout:pg_num: 1 2026-03-09T20:38:47.533 INFO:tasks.cephadm:Setting up client nodes... 2026-03-09T20:38:47.533 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-09T20:38:47.533 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-09T20:38:47.533 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph mgr dump --format=json 2026-03-09T20:38:47.738 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:47.808 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T20:38:47.808 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:47 vm03 ceph-mon[52226]: osdmap e19: 3 total, 3 up, 3 in 2026-03-09T20:38:47.808 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:47 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T20:38:47.808 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:47 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/719812714' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:38:47.808 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:47 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/293400133' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-09T20:38:47.944 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:47 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T20:38:47.944 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:47 vm04 ceph-mon[53250]: osdmap e19: 3 total, 3 up, 3 in 2026-03-09T20:38:47.944 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:47 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T20:38:47.944 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:47 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/719812714' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:38:47.944 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:47 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/293400133' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-09T20:38:47.968 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:47 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T20:38:47.968 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:47 vm08 ceph-mon[54477]: osdmap e19: 3 total, 3 up, 3 in 2026-03-09T20:38:47.968 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:47 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T20:38:47.968 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:47 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/719812714' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:38:47.968 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:47 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/293400133' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-09T20:38:48.036 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:38:48.072 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:38:47 vm03 sudo[62612]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-09T20:38:48.073 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:38:47 vm03 sudo[62612]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T20:38:48.073 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:38:47 vm03 sudo[62612]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T20:38:48.073 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:38:47 vm03 sudo[62612]: pam_unix(sudo:session): session closed for user root 2026-03-09T20:38:48.073 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:47 vm03 sudo[62616]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T20:38:48.073 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:47 vm03 sudo[62616]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T20:38:48.073 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:47 vm03 sudo[62616]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T20:38:48.073 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 sudo[62616]: pam_unix(sudo:session): session closed for user root 2026-03-09T20:38:48.114 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":13,"flags":0,"active_gid":14150,"active_name":"a","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6800","nonce":3921159159},{"type":"v1","addr":"192.168.123.103:6801","nonce":3921159159}]},"active_addr":"192.168.123.103:6801/3921159159","active_change":"2026-03-09T20:37:37.656458+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":24104,"name":"b","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.103:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.103:0","nonce":889556593}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.103:0","nonce":2916604626}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.103:0","nonce":2676531897}]}]} 2026-03-09T20:38:48.116 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-09T20:38:48.116 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-09T20:38:48.116 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph osd dump --format=json 2026-03-09T20:38:48.255 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:38:47 vm04 sudo[58797]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-09T20:38:48.255 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:38:47 vm04 sudo[58797]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T20:38:48.255 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:38:47 vm04 sudo[58797]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T20:38:48.255 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:38:47 vm04 sudo[58797]: pam_unix(sudo:session): session closed for user root 2026-03-09T20:38:48.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 sudo[58801]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T20:38:48.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 sudo[58801]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T20:38:48.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 sudo[58801]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T20:38:48.255 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 sudo[58801]: pam_unix(sudo:session): session closed for user root 2026-03-09T20:38:48.308 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:48.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 sudo[59346]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T20:38:48.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 sudo[59346]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T20:38:48.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 sudo[59346]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T20:38:48.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 sudo[59346]: pam_unix(sudo:session): session closed for user root 2026-03-09T20:38:48.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:38:47 vm08 sudo[59342]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-09T20:38:48.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:38:47 vm08 sudo[59342]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T20:38:48.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:38:47 vm08 sudo[59342]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T20:38:48.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:38:47 vm08 sudo[59342]: pam_unix(sudo:session): session closed for user root 2026-03-09T20:38:48.555 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:38:48.555 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":20,"fsid":"8176208c-1bf7-11f1-86ba-4f626c793dbd","created":"2026-03-09T20:37:15.404226+0000","modified":"2026-03-09T20:38:47.529292+0000","last_up_change":"2026-03-09T20:38:45.396378+0000","last_in_change":"2026-03-09T20:38:31.419551+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T20:38:45.717276+0000","flags":32769,"flags_names":"hashpspool,creating","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"20","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"fe57b656-0d2c-4d80-8c77-0f940eeaf7e5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6802","nonce":351351582},{"type":"v1","addr":"192.168.123.103:6803","nonce":351351582}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6804","nonce":351351582},{"type":"v1","addr":"192.168.123.103:6805","nonce":351351582}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6808","nonce":351351582},{"type":"v1","addr":"192.168.123.103:6809","nonce":351351582}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6806","nonce":351351582},{"type":"v1","addr":"192.168.123.103:6807","nonce":351351582}]},"public_addr":"192.168.123.103:6803/351351582","cluster_addr":"192.168.123.103:6805/351351582","heartbeat_back_addr":"192.168.123.103:6809/351351582","heartbeat_front_addr":"192.168.123.103:6807/351351582","state":["exists","up"]},{"osd":1,"uuid":"d9693d38-ddb7-40f9-bcab-2b2f12cb4552","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":14,"up_thru":19,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":3429737005},{"type":"v1","addr":"192.168.123.104:6801","nonce":3429737005}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":3429737005},{"type":"v1","addr":"192.168.123.104:6803","nonce":3429737005}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":3429737005},{"type":"v1","addr":"192.168.123.104:6807","nonce":3429737005}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":3429737005},{"type":"v1","addr":"192.168.123.104:6805","nonce":3429737005}]},"public_addr":"192.168.123.104:6801/3429737005","cluster_addr":"192.168.123.104:6803/3429737005","heartbeat_back_addr":"192.168.123.104:6807/3429737005","heartbeat_front_addr":"192.168.123.104:6805/3429737005","state":["exists","up"]},{"osd":2,"uuid":"a7e4dc58-4f9e-4949-84ac-fa3c14484ade","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":4157019367},{"type":"v1","addr":"192.168.123.108:6801","nonce":4157019367}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":4157019367},{"type":"v1","addr":"192.168.123.108:6803","nonce":4157019367}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":4157019367},{"type":"v1","addr":"192.168.123.108:6807","nonce":4157019367}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":4157019367},{"type":"v1","addr":"192.168.123.108:6805","nonce":4157019367}]},"public_addr":"192.168.123.108:6801/4157019367","cluster_addr":"192.168.123.108:6803/4157019367","heartbeat_back_addr":"192.168.123.108:6807/4157019367","heartbeat_front_addr":"192.168.123.108:6805/4157019367","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:38:19.762215+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:38:29.781918+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:38:42.680940+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.103:6801/512166571":"2026-03-10T20:37:37.656342+0000","192.168.123.103:0/2604231392":"2026-03-10T20:37:37.656342+0000","192.168.123.103:6800/512166571":"2026-03-10T20:37:37.656342+0000","192.168.123.103:0/3477391265":"2026-03-10T20:37:37.656342+0000","192.168.123.103:0/353200345":"2026-03-10T20:37:37.656342+0000","192.168.123.103:0/362558974":"2026-03-10T20:37:27.651886+0000","192.168.123.103:0/2908564848":"2026-03-10T20:37:27.651886+0000","192.168.123.103:0/3602358308":"2026-03-10T20:37:27.651886+0000","192.168.123.103:6801/3012786188":"2026-03-10T20:37:27.651886+0000","192.168.123.103:6800/3012786188":"2026-03-10T20:37:27.651886+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T20:38:48.635 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-09T20:38:48.635 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph osd dump --format=json 2026-03-09T20:38:48.824 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:48.852 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: osdmap e20: 3 total, 3 up, 3 in 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1439350507' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: pgmap v49: 1 pgs: 1 active+clean; 320 KiB data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:48.853 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:48 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/3801523723' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: osdmap e20: 3 total, 3 up, 3 in 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1439350507' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: pgmap v49: 1 pgs: 1 active+clean; 320 KiB data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:49.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:48 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/3801523723' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:38:49.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: osdmap e20: 3 total, 3 up, 3 in 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1439350507' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:49.069 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T20:38:49.070 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: pgmap v49: 1 pgs: 1 active+clean; 320 KiB data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:38:49.070 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:49.070 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:49.070 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='mgr.14150 192.168.123.103:0/3895172909' entity='mgr.a' 2026-03-09T20:38:49.070 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:48 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/3801523723' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:38:49.099 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:38:49.099 INFO:teuthology.orchestra.run.vm03.stdout:{"epoch":21,"fsid":"8176208c-1bf7-11f1-86ba-4f626c793dbd","created":"2026-03-09T20:37:15.404226+0000","modified":"2026-03-09T20:38:48.590008+0000","last_up_change":"2026-03-09T20:38:45.396378+0000","last_in_change":"2026-03-09T20:38:31.419551+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":8,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":3,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T20:38:45.717276+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"21","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":3,"score_stable":3,"optimal_score":1,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"fe57b656-0d2c-4d80-8c77-0f940eeaf7e5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6802","nonce":351351582},{"type":"v1","addr":"192.168.123.103:6803","nonce":351351582}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6804","nonce":351351582},{"type":"v1","addr":"192.168.123.103:6805","nonce":351351582}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6808","nonce":351351582},{"type":"v1","addr":"192.168.123.103:6809","nonce":351351582}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.103:6806","nonce":351351582},{"type":"v1","addr":"192.168.123.103:6807","nonce":351351582}]},"public_addr":"192.168.123.103:6803/351351582","cluster_addr":"192.168.123.103:6805/351351582","heartbeat_back_addr":"192.168.123.103:6809/351351582","heartbeat_front_addr":"192.168.123.103:6807/351351582","state":["exists","up"]},{"osd":1,"uuid":"d9693d38-ddb7-40f9-bcab-2b2f12cb4552","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":14,"up_thru":19,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":3429737005},{"type":"v1","addr":"192.168.123.104:6801","nonce":3429737005}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":3429737005},{"type":"v1","addr":"192.168.123.104:6803","nonce":3429737005}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":3429737005},{"type":"v1","addr":"192.168.123.104:6807","nonce":3429737005}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":3429737005},{"type":"v1","addr":"192.168.123.104:6805","nonce":3429737005}]},"public_addr":"192.168.123.104:6801/3429737005","cluster_addr":"192.168.123.104:6803/3429737005","heartbeat_back_addr":"192.168.123.104:6807/3429737005","heartbeat_front_addr":"192.168.123.104:6805/3429737005","state":["exists","up"]},{"osd":2,"uuid":"a7e4dc58-4f9e-4949-84ac-fa3c14484ade","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":4157019367},{"type":"v1","addr":"192.168.123.108:6801","nonce":4157019367}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":4157019367},{"type":"v1","addr":"192.168.123.108:6803","nonce":4157019367}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":4157019367},{"type":"v1","addr":"192.168.123.108:6807","nonce":4157019367}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":4157019367},{"type":"v1","addr":"192.168.123.108:6805","nonce":4157019367}]},"public_addr":"192.168.123.108:6801/4157019367","cluster_addr":"192.168.123.108:6803/4157019367","heartbeat_back_addr":"192.168.123.108:6807/4157019367","heartbeat_front_addr":"192.168.123.108:6805/4157019367","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:38:19.762215+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:38:29.781918+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T20:38:42.680940+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.103:6801/512166571":"2026-03-10T20:37:37.656342+0000","192.168.123.103:0/2604231392":"2026-03-10T20:37:37.656342+0000","192.168.123.103:6800/512166571":"2026-03-10T20:37:37.656342+0000","192.168.123.103:0/3477391265":"2026-03-10T20:37:37.656342+0000","192.168.123.103:0/353200345":"2026-03-10T20:37:37.656342+0000","192.168.123.103:0/362558974":"2026-03-10T20:37:27.651886+0000","192.168.123.103:0/2908564848":"2026-03-10T20:37:27.651886+0000","192.168.123.103:0/3602358308":"2026-03-10T20:37:27.651886+0000","192.168.123.103:6801/3012786188":"2026-03-10T20:37:27.651886+0000","192.168.123.103:6800/3012786188":"2026-03-10T20:37:27.651886+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T20:38:49.261 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph tell osd.0 flush_pg_stats 2026-03-09T20:38:49.262 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph tell osd.1 flush_pg_stats 2026-03-09T20:38:49.262 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph tell osd.2 flush_pg_stats 2026-03-09T20:38:49.509 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:49.570 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:49.571 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:49.830 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:49 vm03 ceph-mon[52226]: osdmap e21: 3 total, 3 up, 3 in 2026-03-09T20:38:49.830 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:49 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/4167678828' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:38:49.889 INFO:teuthology.orchestra.run.vm03.stdout:60129542149 2026-03-09T20:38:49.889 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph osd last-stat-seq osd.1 2026-03-09T20:38:49.916 INFO:teuthology.orchestra.run.vm03.stdout:38654705671 2026-03-09T20:38:49.917 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph osd last-stat-seq osd.0 2026-03-09T20:38:49.923 INFO:teuthology.orchestra.run.vm03.stdout:77309411330 2026-03-09T20:38:49.923 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph osd last-stat-seq osd.2 2026-03-09T20:38:50.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:49 vm04 ceph-mon[53250]: osdmap e21: 3 total, 3 up, 3 in 2026-03-09T20:38:50.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:49 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/4167678828' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:38:50.067 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:49 vm08 ceph-mon[54477]: osdmap e21: 3 total, 3 up, 3 in 2026-03-09T20:38:50.067 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:49 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/4167678828' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T20:38:50.122 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:50.285 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:50.323 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:50.451 INFO:teuthology.orchestra.run.vm03.stdout:60129542148 2026-03-09T20:38:50.583 INFO:tasks.cephadm.ceph_manager.ceph:need seq 60129542149 got 60129542148 for osd.1 2026-03-09T20:38:50.640 INFO:teuthology.orchestra.run.vm03.stdout:77309411330 2026-03-09T20:38:50.654 INFO:teuthology.orchestra.run.vm03.stdout:38654705671 2026-03-09T20:38:50.673 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:50 vm03 ceph-mon[52226]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 4.5 KiB/s rd, 131 KiB/s wr, 12 op/s 2026-03-09T20:38:50.673 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:50 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/2929391599' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T20:38:50.720 INFO:tasks.cephadm.ceph_manager.ceph:need seq 38654705671 got 38654705671 for osd.0 2026-03-09T20:38:50.720 DEBUG:teuthology.parallel:result is None 2026-03-09T20:38:50.751 INFO:tasks.cephadm.ceph_manager.ceph:need seq 77309411330 got 77309411330 for osd.2 2026-03-09T20:38:50.751 DEBUG:teuthology.parallel:result is None 2026-03-09T20:38:51.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:50 vm04 ceph-mon[53250]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 4.5 KiB/s rd, 131 KiB/s wr, 12 op/s 2026-03-09T20:38:51.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:50 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/2929391599' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T20:38:51.067 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:50 vm08 ceph-mon[54477]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 4.5 KiB/s rd, 131 KiB/s wr, 12 op/s 2026-03-09T20:38:51.067 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:50 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/2929391599' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T20:38:51.583 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph osd last-stat-seq osd.1 2026-03-09T20:38:51.772 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:51.888 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:51 vm03 ceph-mon[52226]: mgrmap e14: a(active, since 72s), standbys: b 2026-03-09T20:38:51.888 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:51 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/4091530187' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T20:38:51.888 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:51 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/2903461811' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T20:38:52.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:51 vm04 ceph-mon[53250]: mgrmap e14: a(active, since 72s), standbys: b 2026-03-09T20:38:52.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:51 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/4091530187' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T20:38:52.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:51 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/2903461811' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T20:38:52.007 INFO:teuthology.orchestra.run.vm03.stdout:60129542149 2026-03-09T20:38:52.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:51 vm08 ceph-mon[54477]: mgrmap e14: a(active, since 72s), standbys: b 2026-03-09T20:38:52.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:51 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/4091530187' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T20:38:52.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:51 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/2903461811' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T20:38:52.081 INFO:tasks.cephadm.ceph_manager.ceph:need seq 60129542149 got 60129542149 for osd.1 2026-03-09T20:38:52.081 DEBUG:teuthology.parallel:result is None 2026-03-09T20:38:52.081 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-09T20:38:52.081 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph pg dump --format=json 2026-03-09T20:38:52.261 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:52.492 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:38:52.492 INFO:teuthology.orchestra.run.vm03.stderr:dumped all 2026-03-09T20:38:52.552 INFO:teuthology.orchestra.run.vm03.stdout:{"pg_ready":true,"pg_map":{"version":52,"stamp":"2026-03-09T20:38:52.239245+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":3,"num_per_pool_osds":3,"num_per_pool_omap_osds":3,"kb":62902272,"kb_used":82784,"kb_used_data":1884,"kb_used_omap":4,"kb_used_meta":80443,"kb_avail":62819488,"statfs":{"total":64411926528,"available":64327155712,"internally_reserved":0,"allocated":1929216,"data_stored":1540872,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":4770,"internal_metadata":82373982},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":131088,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":10,"num_read_kb":9,"num_write":16,"num_write_kb":262,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.000646"},"pg_stats":[{"pgid":"1.0","version":"20'32","reported_seq":57,"reported_epoch":21,"state":"active+clean","last_fresh":"2026-03-09T20:38:48.602861+0000","last_change":"2026-03-09T20:38:47.831139+0000","last_active":"2026-03-09T20:38:48.602861+0000","last_peered":"2026-03-09T20:38:48.602861+0000","last_clean":"2026-03-09T20:38:48.602861+0000","last_became_active":"2026-03-09T20:38:47.830882+0000","last_became_peered":"2026-03-09T20:38:47.830882+0000","last_unstale":"2026-03-09T20:38:48.602861+0000","last_undegraded":"2026-03-09T20:38:48.602861+0000","last_fullsized":"2026-03-09T20:38:48.602861+0000","mapping_epoch":19,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":20,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T20:38:46.520776+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T20:38:46.520776+0000","last_clean_scrub_stamp":"2026-03-09T20:38:46.520776+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T07:23:14.165756+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,0],"acting":[1,2,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":2,"up_from":18,"seq":77309411331,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27592,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939832,"statfs":{"total":21470642176,"available":21442387968,"internally_reserved":0,"allocated":643072,"data_stored":513624,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":14,"seq":60129542149,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27592,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939832,"statfs":{"total":21470642176,"available":21442387968,"internally_reserved":0,"allocated":643072,"data_stored":513624,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[]},{"osd":0,"up_from":9,"seq":38654705672,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27600,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939824,"statfs":{"total":21470642176,"available":21442379776,"internally_reserved":0,"allocated":643072,"data_stored":513624,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T20:38:52.553 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph pg dump --format=json 2026-03-09T20:38:52.733 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:52.793 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:52 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/3983096744' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T20:38:52.793 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:52 vm03 ceph-mon[52226]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 2.2 KiB/s rd, 65 KiB/s wr, 6 op/s 2026-03-09T20:38:53.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:52 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/3983096744' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T20:38:53.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:52 vm04 ceph-mon[53250]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 2.2 KiB/s rd, 65 KiB/s wr, 6 op/s 2026-03-09T20:38:53.012 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:38:53.012 INFO:teuthology.orchestra.run.vm03.stderr:dumped all 2026-03-09T20:38:53.067 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:52 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/3983096744' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T20:38:53.067 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:52 vm08 ceph-mon[54477]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 2.2 KiB/s rd, 65 KiB/s wr, 6 op/s 2026-03-09T20:38:53.079 INFO:teuthology.orchestra.run.vm03.stdout:{"pg_ready":true,"pg_map":{"version":52,"stamp":"2026-03-09T20:38:52.239245+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":3,"num_per_pool_osds":3,"num_per_pool_omap_osds":3,"kb":62902272,"kb_used":82784,"kb_used_data":1884,"kb_used_omap":4,"kb_used_meta":80443,"kb_avail":62819488,"statfs":{"total":64411926528,"available":64327155712,"internally_reserved":0,"allocated":1929216,"data_stored":1540872,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":4770,"internal_metadata":82373982},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":131088,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":10,"num_read_kb":9,"num_write":16,"num_write_kb":262,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.000646"},"pg_stats":[{"pgid":"1.0","version":"20'32","reported_seq":57,"reported_epoch":21,"state":"active+clean","last_fresh":"2026-03-09T20:38:48.602861+0000","last_change":"2026-03-09T20:38:47.831139+0000","last_active":"2026-03-09T20:38:48.602861+0000","last_peered":"2026-03-09T20:38:48.602861+0000","last_clean":"2026-03-09T20:38:48.602861+0000","last_became_active":"2026-03-09T20:38:47.830882+0000","last_became_peered":"2026-03-09T20:38:47.830882+0000","last_unstale":"2026-03-09T20:38:48.602861+0000","last_undegraded":"2026-03-09T20:38:48.602861+0000","last_fullsized":"2026-03-09T20:38:48.602861+0000","mapping_epoch":19,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":20,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T20:38:46.520776+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T20:38:46.520776+0000","last_clean_scrub_stamp":"2026-03-09T20:38:46.520776+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T07:23:14.165756+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,2,0],"acting":[1,2,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":2,"up_from":18,"seq":77309411331,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27592,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939832,"statfs":{"total":21470642176,"available":21442387968,"internally_reserved":0,"allocated":643072,"data_stored":513624,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":14,"seq":60129542149,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27592,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939832,"statfs":{"total":21470642176,"available":21442387968,"internally_reserved":0,"allocated":643072,"data_stored":513624,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[]},{"osd":0,"up_from":9,"seq":38654705672,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27600,"kb_used_data":628,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939824,"statfs":{"total":21470642176,"available":21442379776,"internally_reserved":0,"allocated":643072,"data_stored":513624,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T20:38:53.079 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-09T20:38:53.079 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-09T20:38:53.079 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-09T20:38:53.079 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph health --format=json 2026-03-09T20:38:53.256 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:53.512 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:38:53.512 INFO:teuthology.orchestra.run.vm03.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-09T20:38:53.590 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-09T20:38:53.590 INFO:tasks.cephadm:Setup complete, yielding 2026-03-09T20:38:53.590 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-09T20:38:53.592 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm03.local 2026-03-09T20:38:53.592 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- bash -c 'ceph mgr module enable rgw' 2026-03-09T20:38:53.782 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:53.810 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:53 vm03 ceph-mon[52226]: from='client.14391 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:38:53.810 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:53 vm03 ceph-mon[52226]: from='client.24244 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:38:53.810 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:53 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/4080238589' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T20:38:54.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:53 vm04 ceph-mon[53250]: from='client.14391 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:38:54.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:53 vm04 ceph-mon[53250]: from='client.24244 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:38:54.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:53 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/4080238589' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T20:38:54.067 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:53 vm08 ceph-mon[54477]: from='client.14391 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:38:54.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:53 vm08 ceph-mon[54477]: from='client.24244 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T20:38:54.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:53 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/4080238589' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T20:38:54.725 INFO:teuthology.run_tasks:Running task rgw_module.apply... 2026-03-09T20:38:54.728 INFO:tasks.rgw_module:Applying spec(s): rgw_realm: myrealm1 rgw_zone: myzone1 rgw_zonegroup: myzonegroup1 spec: rgw_frontend_port: 5500 2026-03-09T20:38:54.728 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph rgw realm bootstrap -i - 2026-03-09T20:38:54.895 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:54 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1830070615' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "rgw"}]: dispatch 2026-03-09T20:38:54.895 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:54 vm03 ceph-mon[52226]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 1.5 KiB/s rd, 44 KiB/s wr, 4 op/s 2026-03-09T20:38:54.895 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:54 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: ignoring --setuser ceph since I am not root 2026-03-09T20:38:54.895 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:54 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: ignoring --setgroup ceph since I am not root 2026-03-09T20:38:54.895 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:54 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:54.790+0000 7f20f917c140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T20:38:54.895 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:54 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:54.832+0000 7f20f917c140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T20:38:54.920 INFO:teuthology.orchestra.run.vm03.stdout:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:38:55.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:54 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1830070615' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "rgw"}]: dispatch 2026-03-09T20:38:55.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:54 vm04 ceph-mon[53250]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 1.5 KiB/s rd, 44 KiB/s wr, 4 op/s 2026-03-09T20:38:55.005 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:54 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: ignoring --setuser ceph since I am not root 2026-03-09T20:38:55.005 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:54 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: ignoring --setgroup ceph since I am not root 2026-03-09T20:38:55.005 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:54 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:54.784+0000 7ff30d78c140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T20:38:55.005 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:54 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:54.830+0000 7ff30d78c140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T20:38:55.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:54 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1830070615' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "rgw"}]: dispatch 2026-03-09T20:38:55.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:54 vm08 ceph-mon[54477]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 1.5 KiB/s rd, 44 KiB/s wr, 4 op/s 2026-03-09T20:38:55.637 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:55 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:55.304+0000 7f20f917c140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T20:38:55.637 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:55 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:55.297+0000 7ff30d78c140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T20:38:55.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:55 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1830070615' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "rgw"}]': finished 2026-03-09T20:38:55.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:55 vm03 ceph-mon[52226]: mgrmap e15: a(active, since 76s), standbys: b 2026-03-09T20:38:55.894 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:55 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:55.656+0000 7f20f917c140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T20:38:55.894 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:55 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T20:38:55.894 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:55 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T20:38:55.894 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:55 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: from numpy import show_config as show_numpy_config 2026-03-09T20:38:55.894 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:55 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:55.753+0000 7f20f917c140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T20:38:55.894 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:55 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:55.795+0000 7f20f917c140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T20:38:55.894 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:55 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:55.876+0000 7f20f917c140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T20:38:56.005 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:55 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:55.637+0000 7ff30d78c140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T20:38:56.005 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:55 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T20:38:56.005 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:55 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T20:38:56.005 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:55 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: from numpy import show_config as show_numpy_config 2026-03-09T20:38:56.005 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:55 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:55.734+0000 7ff30d78c140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T20:38:56.005 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:55 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:55.775+0000 7ff30d78c140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T20:38:56.005 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:55 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:55.853+0000 7ff30d78c140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T20:38:56.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:55 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1830070615' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "rgw"}]': finished 2026-03-09T20:38:56.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:55 vm04 ceph-mon[53250]: mgrmap e15: a(active, since 76s), standbys: b 2026-03-09T20:38:56.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:55 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1830070615' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "rgw"}]': finished 2026-03-09T20:38:56.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:55 vm08 ceph-mon[54477]: mgrmap e15: a(active, since 76s), standbys: b 2026-03-09T20:38:56.679 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:56 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:56.409+0000 7ff30d78c140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T20:38:56.679 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:56 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:56.541+0000 7ff30d78c140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:38:56.679 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:56 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:56.589+0000 7ff30d78c140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T20:38:56.679 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:56 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:56.631+0000 7ff30d78c140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T20:38:56.686 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:56 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:56.431+0000 7f20f917c140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T20:38:56.686 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:56 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:56.554+0000 7f20f917c140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:38:56.686 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:56 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:56.599+0000 7f20f917c140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T20:38:56.686 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:56 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:56.638+0000 7f20f917c140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T20:38:56.686 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:56 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:56.685+0000 7f20f917c140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T20:38:56.962 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:56 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:56.677+0000 7ff30d78c140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T20:38:56.962 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:56 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:56.719+0000 7ff30d78c140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T20:38:56.962 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:56 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:56.909+0000 7ff30d78c140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T20:38:56.978 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:56 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:56.728+0000 7f20f917c140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T20:38:56.978 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:56 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:56.920+0000 7f20f917c140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T20:38:57.255 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:56 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:56.961+0000 7ff30d78c140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T20:38:57.255 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:57 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:57.200+0000 7ff30d78c140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T20:38:57.394 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:56 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:56.977+0000 7f20f917c140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T20:38:57.394 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:57 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:57.216+0000 7f20f917c140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T20:38:57.755 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:57 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:57.496+0000 7ff30d78c140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T20:38:57.755 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:57 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:57.534+0000 7ff30d78c140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T20:38:57.755 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:57 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:57.577+0000 7ff30d78c140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T20:38:57.755 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:57 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:57.655+0000 7ff30d78c140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T20:38:57.755 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:57 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:57.693+0000 7ff30d78c140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T20:38:57.791 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:57 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:57.507+0000 7f20f917c140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T20:38:57.791 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:57 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:57.545+0000 7f20f917c140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T20:38:57.791 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:57 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:57.590+0000 7f20f917c140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T20:38:57.791 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:57 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:57.670+0000 7f20f917c140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T20:38:57.791 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:57 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:57.708+0000 7f20f917c140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T20:38:58.046 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:57 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:57.773+0000 7ff30d78c140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T20:38:58.046 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:57 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:57.898+0000 7ff30d78c140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:38:58.050 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:57 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:57.789+0000 7f20f917c140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T20:38:58.050 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:57 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:57.909+0000 7f20f917c140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:58.048+0000 7f20f917c140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a[52436]: 2026-03-09T20:38:58.087+0000 7f20f917c140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: Standby manager daemon b restarted 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: Standby manager daemon b started 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: Active manager daemon a restarted 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: Activating manager daemon a 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.? 192.168.123.104:0/1082354810' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.? 192.168.123.104:0/1082354810' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.? 192.168.123.104:0/1082354810' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.? 192.168.123.104:0/1082354810' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: osdmap e22: 3 total, 3 up, 3 in 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: mgrmap e16: a(active, starting, since 0.0127025s), standbys: b 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: Manager daemon a is now available 2026-03-09T20:38:58.394 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:58 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:58.505 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:58.045+0000 7ff30d78c140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T20:38:58.505 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-b[54694]: 2026-03-09T20:38:58.087+0000 7ff30d78c140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T20:38:58.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: Standby manager daemon b restarted 2026-03-09T20:38:58.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: Standby manager daemon b started 2026-03-09T20:38:58.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: Active manager daemon a restarted 2026-03-09T20:38:58.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: Activating manager daemon a 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.? 192.168.123.104:0/1082354810' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.? 192.168.123.104:0/1082354810' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.? 192.168.123.104:0/1082354810' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.? 192.168.123.104:0/1082354810' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: osdmap e22: 3 total, 3 up, 3 in 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: mgrmap e16: a(active, starting, since 0.0127025s), standbys: b 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: Manager daemon a is now available 2026-03-09T20:38:58.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:58 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:58.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: Standby manager daemon b restarted 2026-03-09T20:38:58.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: Standby manager daemon b started 2026-03-09T20:38:58.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: Active manager daemon a restarted 2026-03-09T20:38:58.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: Activating manager daemon a 2026-03-09T20:38:58.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.? 192.168.123.104:0/1082354810' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T20:38:58.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.? 192.168.123.104:0/1082354810' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T20:38:58.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.? 192.168.123.104:0/1082354810' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T20:38:58.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.? 192.168.123.104:0/1082354810' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T20:38:58.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: osdmap e22: 3 total, 3 up, 3 in 2026-03-09T20:38:58.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: mgrmap e16: a(active, starting, since 0.0127025s), standbys: b 2026-03-09T20:38:58.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T20:38:58.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T20:38:58.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T20:38:58.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T20:38:58.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T20:38:58.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T20:38:58.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T20:38:58.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T20:38:58.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T20:38:58.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T20:38:58.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T20:38:58.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: Manager daemon a is now available 2026-03-09T20:38:58.569 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:58 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd='[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]': finished 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:38:59 vm04 ceph-mon[53250]: mgrmap e17: a(active, since 1.02269s), standbys: b 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd='[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]': finished 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:38:59 vm08 ceph-mon[54477]: mgrmap e17: a(active, since 1.02269s), standbys: b 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm03", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd='[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]': finished 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: Updating vm03:/etc/ceph/ceph.conf 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.conf 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: Updating vm03:/etc/ceph/ceph.client.admin.keyring 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:38:59.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:38:59 vm03 ceph-mon[52226]: mgrmap e17: a(active, since 1.02269s), standbys: b 2026-03-09T20:39:00.249 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:00 vm03 ceph-mon[52226]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:39:00.249 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:00 vm03 ceph-mon[52226]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:39:00.249 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:00 vm03 ceph-mon[52226]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:39:00.249 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:00 vm03 ceph-mon[52226]: [09/Mar/2026:20:38:59] ENGINE Bus STARTING 2026-03-09T20:39:00.249 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:00 vm03 ceph-mon[52226]: from='client.24259 -' entity='client.admin' cmd=[{"prefix": "rgw realm bootstrap", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:39:00.249 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:00 vm03 ceph-mon[52226]: pgmap v3: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:39:00.249 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:00 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-09T20:39:00.249 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:00 vm03 ceph-mon[52226]: [09/Mar/2026:20:38:59] ENGINE Serving on http://192.168.123.103:8765 2026-03-09T20:39:00.249 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:00 vm03 ceph-mon[52226]: [09/Mar/2026:20:38:59] ENGINE Serving on https://192.168.123.103:7151 2026-03-09T20:39:00.249 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:00 vm03 ceph-mon[52226]: [09/Mar/2026:20:38:59] ENGINE Bus STARTED 2026-03-09T20:39:00.249 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:00 vm03 ceph-mon[52226]: [09/Mar/2026:20:38:59] ENGINE Client ('192.168.123.103', 46600) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T20:39:00.249 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:00 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm04", "caps": []}]: dispatch 2026-03-09T20:39:00.249 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:00 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:00.470 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:00 vm08 ceph-mon[54477]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:39:00.471 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:00 vm08 ceph-mon[54477]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:39:00.471 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:00 vm08 ceph-mon[54477]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:39:00.471 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:00 vm08 ceph-mon[54477]: [09/Mar/2026:20:38:59] ENGINE Bus STARTING 2026-03-09T20:39:00.471 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:00 vm08 ceph-mon[54477]: from='client.24259 -' entity='client.admin' cmd=[{"prefix": "rgw realm bootstrap", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:39:00.471 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:00 vm08 ceph-mon[54477]: pgmap v3: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:39:00.471 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:00 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-09T20:39:00.471 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:00 vm08 ceph-mon[54477]: [09/Mar/2026:20:38:59] ENGINE Serving on http://192.168.123.103:8765 2026-03-09T20:39:00.471 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:00 vm08 ceph-mon[54477]: [09/Mar/2026:20:38:59] ENGINE Serving on https://192.168.123.103:7151 2026-03-09T20:39:00.471 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:00 vm08 ceph-mon[54477]: [09/Mar/2026:20:38:59] ENGINE Bus STARTED 2026-03-09T20:39:00.471 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:00 vm08 ceph-mon[54477]: [09/Mar/2026:20:38:59] ENGINE Client ('192.168.123.103', 46600) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T20:39:00.471 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:00 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm04", "caps": []}]: dispatch 2026-03-09T20:39:00.471 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:00 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:00.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:00 vm04 ceph-mon[53250]: Updating vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:39:00.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:00 vm04 ceph-mon[53250]: Updating vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:39:00.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:00 vm04 ceph-mon[53250]: Updating vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/config/ceph.client.admin.keyring 2026-03-09T20:39:00.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:00 vm04 ceph-mon[53250]: [09/Mar/2026:20:38:59] ENGINE Bus STARTING 2026-03-09T20:39:00.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:00 vm04 ceph-mon[53250]: from='client.24259 -' entity='client.admin' cmd=[{"prefix": "rgw realm bootstrap", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:39:00.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:00 vm04 ceph-mon[53250]: pgmap v3: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:39:00.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:00 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm03", "caps": []}]: dispatch 2026-03-09T20:39:00.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:00 vm04 ceph-mon[53250]: [09/Mar/2026:20:38:59] ENGINE Serving on http://192.168.123.103:8765 2026-03-09T20:39:00.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:00 vm04 ceph-mon[53250]: [09/Mar/2026:20:38:59] ENGINE Serving on https://192.168.123.103:7151 2026-03-09T20:39:00.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:00 vm04 ceph-mon[53250]: [09/Mar/2026:20:38:59] ENGINE Bus STARTED 2026-03-09T20:39:00.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:00 vm04 ceph-mon[53250]: [09/Mar/2026:20:38:59] ENGINE Client ('192.168.123.103', 46600) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T20:39:00.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:00 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm04", "caps": []}]: dispatch 2026-03-09T20:39:00.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:00 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:01.289 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:01 vm03 ceph-mon[52226]: pgmap v4: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:39:01.289 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:01 vm03 ceph-mon[52226]: osdmap e23: 3 total, 3 up, 3 in 2026-03-09T20:39:01.289 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:01 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/2128200231' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-09T20:39:01.289 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:01 vm03 ceph-mon[52226]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-09T20:39:01.289 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:01 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm08", "caps": []}]: dispatch 2026-03-09T20:39:01.289 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:01 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:01.289 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:01 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:01.289 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:01 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:01.289 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:01 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:01.289 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:01 vm03 ceph-mon[52226]: mgrmap e18: a(active, since 2s), standbys: b 2026-03-09T20:39:01.289 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:01 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:01.289 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:01 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm04", "caps": []}]: dispatch 2026-03-09T20:39:01.289 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:01 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm08", "caps": []}]: dispatch 2026-03-09T20:39:01.391 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:01 vm04 ceph-mon[53250]: pgmap v4: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:39:01.391 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:01 vm04 ceph-mon[53250]: osdmap e23: 3 total, 3 up, 3 in 2026-03-09T20:39:01.391 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:01 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/2128200231' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-09T20:39:01.391 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:01 vm04 ceph-mon[53250]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-09T20:39:01.391 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:01 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm08", "caps": []}]: dispatch 2026-03-09T20:39:01.391 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:01 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:01.391 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:01 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:01.391 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:01 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:01.391 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:01 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:01.391 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:01 vm04 ceph-mon[53250]: mgrmap e18: a(active, since 2s), standbys: b 2026-03-09T20:39:01.391 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:01 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:01.391 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:01 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm04", "caps": []}]: dispatch 2026-03-09T20:39:01.391 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:01 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm08", "caps": []}]: dispatch 2026-03-09T20:39:01.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:01 vm08 ceph-mon[54477]: pgmap v4: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:39:01.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:01 vm08 ceph-mon[54477]: osdmap e23: 3 total, 3 up, 3 in 2026-03-09T20:39:01.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:01 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/2128200231' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-09T20:39:01.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:01 vm08 ceph-mon[54477]: from='client.? ' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-09T20:39:01.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:01 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm08", "caps": []}]: dispatch 2026-03-09T20:39:01.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:01 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:01.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:01 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:01.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:01 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:01.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:01 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:01.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:01 vm08 ceph-mon[54477]: mgrmap e18: a(active, since 2s), standbys: b 2026-03-09T20:39:01.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:01 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:01.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:01 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm04", "caps": []}]: dispatch 2026-03-09T20:39:01.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:01 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm08", "caps": []}]: dispatch 2026-03-09T20:39:02.504 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-09T20:39:02.504 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: osdmap e24: 3 total, 3 up, 3 in 2026-03-09T20:39:02.504 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.504 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.504 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.504 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.504 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.504 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.504 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.504 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.504 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.504 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.504 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.504 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:02 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: osdmap e24: 3 total, 3 up, 3 in 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:02 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='client.? ' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-09T20:39:02.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: osdmap e24: 3 total, 3 up, 3 in 2026-03-09T20:39:02.567 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:02.568 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:02 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:03 vm08 ceph-mon[54477]: pgmap v7: 33 pgs: 32 unknown, 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:39:03.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:03 vm08 ceph-mon[54477]: osdmap e25: 3 total, 3 up, 3 in 2026-03-09T20:39:03.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:03 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:03 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:03 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:03 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:03 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:03 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:03 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:03 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.318 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:03 vm08 ceph-mon[54477]: mgrmap e19: a(active, since 5s), standbys: b 2026-03-09T20:39:03.458 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:03 vm04 ceph-mon[53250]: pgmap v7: 33 pgs: 32 unknown, 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:39:03.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:03 vm04 ceph-mon[53250]: osdmap e25: 3 total, 3 up, 3 in 2026-03-09T20:39:03.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:03 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:03 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:03 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:03 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:03 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:03 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:03 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:03 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:03 vm04 ceph-mon[53250]: mgrmap e19: a(active, since 5s), standbys: b 2026-03-09T20:39:03.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:03 vm03 ceph-mon[52226]: pgmap v7: 33 pgs: 32 unknown, 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:39:03.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:03 vm03 ceph-mon[52226]: osdmap e25: 3 total, 3 up, 3 in 2026-03-09T20:39:03.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:03 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:03 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:03 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:03 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:03 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:03 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:03 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:03 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:03.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:03 vm03 ceph-mon[52226]: mgrmap e19: a(active, since 5s), standbys: b 2026-03-09T20:39:04.371 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:04.371 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:04.371 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: osdmap e26: 3 total, 3 up, 3 in 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]: dispatch 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.372 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:04 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.396 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:04.396 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:04.396 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.396 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:39:04.396 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.396 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.396 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.396 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: osdmap e26: 3 total, 3 up, 3 in 2026-03-09T20:39:04.396 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]: dispatch 2026-03-09T20:39:04.396 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.396 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.396 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.396 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.396 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.397 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.397 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.397 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.397 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:04 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: osdmap e26: 3 total, 3 up, 3 in 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]: dispatch 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.645 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.645 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:04.645 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:04 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: pgmap v10: 65 pgs: 32 creating+peering, 33 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 51 KiB/s rd, 3.5 KiB/s wr, 31 op/s 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: Detected new or changed devices on vm08 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]': finished 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: osdmap e27: 3 total, 3 up, 3 in 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.239 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:05 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.432 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: pgmap v10: 65 pgs: 32 creating+peering, 33 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 51 KiB/s rd, 3.5 KiB/s wr, 31 op/s 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: Detected new or changed devices on vm08 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]': finished 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: osdmap e27: 3 total, 3 up, 3 in 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.433 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:05 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: pgmap v10: 65 pgs: 32 creating+peering, 33 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 51 KiB/s rd, 3.5 KiB/s wr, 31 op/s 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: Detected new or changed devices on vm08 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.log","app": "rgw"}]': finished 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: osdmap e27: 3 total, 3 up, 3 in 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:05.644 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:05 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: osdmap e28: 3 total, 3 up, 3 in 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]: dispatch 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:06 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: osdmap e28: 3 total, 3 up, 3 in 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]: dispatch 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:06.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:06 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: osdmap e28: 3 total, 3 up, 3 in 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]: dispatch 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:07.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:06 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:07.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:07 vm08 ceph-mon[54477]: pgmap v13: 97 pgs: 32 unknown, 32 creating+peering, 33 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 51 KiB/s rd, 3.5 KiB/s wr, 31 op/s 2026-03-09T20:39:07.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:07 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]': finished 2026-03-09T20:39:07.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:07 vm08 ceph-mon[54477]: osdmap e29: 3 total, 3 up, 3 in 2026-03-09T20:39:07.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:07 vm08 ceph-mon[54477]: osdmap e30: 3 total, 3 up, 3 in 2026-03-09T20:39:07.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:07 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]: dispatch 2026-03-09T20:39:07.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:07 vm03 ceph-mon[52226]: pgmap v13: 97 pgs: 32 unknown, 32 creating+peering, 33 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 51 KiB/s rd, 3.5 KiB/s wr, 31 op/s 2026-03-09T20:39:07.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:07 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]': finished 2026-03-09T20:39:07.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:07 vm03 ceph-mon[52226]: osdmap e29: 3 total, 3 up, 3 in 2026-03-09T20:39:07.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:07 vm03 ceph-mon[52226]: osdmap e30: 3 total, 3 up, 3 in 2026-03-09T20:39:07.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:07 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]: dispatch 2026-03-09T20:39:08.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:07 vm04 ceph-mon[53250]: pgmap v13: 97 pgs: 32 unknown, 32 creating+peering, 33 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 51 KiB/s rd, 3.5 KiB/s wr, 31 op/s 2026-03-09T20:39:08.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:07 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.control","app": "rgw"}]': finished 2026-03-09T20:39:08.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:07 vm04 ceph-mon[53250]: osdmap e29: 3 total, 3 up, 3 in 2026-03-09T20:39:08.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:07 vm04 ceph-mon[53250]: osdmap e30: 3 total, 3 up, 3 in 2026-03-09T20:39:08.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:07 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd=[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]: dispatch 2026-03-09T20:39:09.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:09 vm08 ceph-mon[54477]: pgmap v16: 129 pgs: 64 unknown, 32 creating+peering, 33 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:39:09.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:09 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]': finished 2026-03-09T20:39:09.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:09 vm08 ceph-mon[54477]: osdmap e31: 3 total, 3 up, 3 in 2026-03-09T20:39:09.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:09 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd=[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-09T20:39:09.860 INFO:teuthology.orchestra.run.vm03.stdout:Realm(s) created correctly. Please, use 'ceph rgw realm tokens' to get the token. 2026-03-09T20:39:09.860 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:09 vm03 ceph-mon[52226]: pgmap v16: 129 pgs: 64 unknown, 32 creating+peering, 33 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:39:09.860 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:09 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]': finished 2026-03-09T20:39:09.861 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:09 vm03 ceph-mon[52226]: osdmap e31: 3 total, 3 up, 3 in 2026-03-09T20:39:09.861 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:09 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd=[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-09T20:39:09.914 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-09T20:39:09.916 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm03.local 2026-03-09T20:39:09.917 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- bash -c 'set -e 2026-03-09T20:39:09.917 DEBUG:teuthology.orchestra.run.vm03:> set -x 2026-03-09T20:39:09.917 DEBUG:teuthology.orchestra.run.vm03:> while true; do TOKEN=$(ceph rgw realm tokens | jq -r '"'"'.[0].token'"'"'); echo $TOKEN; if [ "$TOKEN" != "master zone has no endpoint" ]; then break; fi; sleep 5; done 2026-03-09T20:39:09.917 DEBUG:teuthology.orchestra.run.vm03:> TOKENS=$(ceph rgw realm tokens) 2026-03-09T20:39:09.917 DEBUG:teuthology.orchestra.run.vm03:> echo $TOKENS | jq --exit-status '"'"'.[0].realm == "myrealm1"'"'"' 2026-03-09T20:39:09.917 DEBUG:teuthology.orchestra.run.vm03:> echo $TOKENS | jq --exit-status '"'"'.[0].token'"'"' 2026-03-09T20:39:09.917 DEBUG:teuthology.orchestra.run.vm03:> TOKEN_JSON=$(ceph rgw realm tokens | jq -r '"'"'.[0].token'"'"' | base64 --decode) 2026-03-09T20:39:09.917 DEBUG:teuthology.orchestra.run.vm03:> echo $TOKEN_JSON | jq --exit-status '"'"'.realm_name == "myrealm1"'"'"' 2026-03-09T20:39:09.917 DEBUG:teuthology.orchestra.run.vm03:> echo $TOKEN_JSON | jq --exit-status '"'"'.endpoint | test("http://.+:\\d+")'"'"' 2026-03-09T20:39:09.917 DEBUG:teuthology.orchestra.run.vm03:> echo $TOKEN_JSON | jq --exit-status '"'"'.realm_id | test("^[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}$")'"'"' 2026-03-09T20:39:09.917 DEBUG:teuthology.orchestra.run.vm03:> echo $TOKEN_JSON | jq --exit-status '"'"'.access_key'"'"' 2026-03-09T20:39:09.917 DEBUG:teuthology.orchestra.run.vm03:> echo $TOKEN_JSON | jq --exit-status '"'"'.secret'"'"' 2026-03-09T20:39:09.917 DEBUG:teuthology.orchestra.run.vm03:> ' 2026-03-09T20:39:10.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:09 vm04 ceph-mon[53250]: pgmap v16: 129 pgs: 64 unknown, 32 creating+peering, 33 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T20:39:10.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:09 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd='[{"prefix": "osd pool application enable","pool": "myzone1.rgw.meta","app": "rgw"}]': finished 2026-03-09T20:39:10.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:09 vm04 ceph-mon[53250]: osdmap e31: 3 total, 3 up, 3 in 2026-03-09T20:39:10.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:09 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd=[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-09T20:39:10.089 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:39:10.175 INFO:teuthology.orchestra.run.vm03.stderr:+ true 2026-03-09T20:39:10.175 INFO:teuthology.orchestra.run.vm03.stderr:++ ceph rgw realm tokens 2026-03-09T20:39:10.177 INFO:teuthology.orchestra.run.vm03.stderr:++ jq -r '.[0].token' 2026-03-09T20:39:10.748 INFO:teuthology.orchestra.run.vm03.stderr:+ TOKEN='master zone has no endpoint' 2026-03-09T20:39:10.749 INFO:teuthology.orchestra.run.vm03.stderr:+ echo master zone has no endpoint 2026-03-09T20:39:10.749 INFO:teuthology.orchestra.run.vm03.stdout:master zone has no endpoint 2026-03-09T20:39:10.749 INFO:teuthology.orchestra.run.vm03.stderr:+ '[' 'master zone has no endpoint' '!=' 'master zone has no endpoint' ']' 2026-03-09T20:39:10.749 INFO:teuthology.orchestra.run.vm03.stderr:+ sleep 5 2026-03-09T20:39:10.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:10 vm04 ceph-mon[53250]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd='[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-09T20:39:10.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:10 vm04 ceph-mon[53250]: osdmap e32: 3 total, 3 up, 3 in 2026-03-09T20:39:10.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:10 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:10 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:10.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:10 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:10.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:10 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:10.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:10 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:10 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:10 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:10 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:10 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm04.knyogr", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-09T20:39:10.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:10 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm04.knyogr", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-09T20:39:10.756 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:10 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.756 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:10 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:10.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:10 vm08 ceph-mon[54477]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd='[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-09T20:39:10.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:10 vm08 ceph-mon[54477]: osdmap e32: 3 total, 3 up, 3 in 2026-03-09T20:39:10.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:10 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:10 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:10.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:10 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:10.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:10 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:10.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:10 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:10 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:10 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:10 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:10 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm04.knyogr", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-09T20:39:10.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:10 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm04.knyogr", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-09T20:39:10.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:10 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.818 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:10 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:10.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:10 vm03 ceph-mon[52226]: from='client.? 192.168.123.103:0/1491757464' entity='mgr.a' cmd='[{"prefix": "osd pool set", "pool": "myzone1.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-09T20:39:10.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:10 vm03 ceph-mon[52226]: osdmap e32: 3 total, 3 up, 3 in 2026-03-09T20:39:10.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:10 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:10 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:10.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:10 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:10.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:10 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:10.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:10 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:10 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:10 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:10 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:10 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm04.knyogr", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-09T20:39:10.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:10 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm04.knyogr", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-09T20:39:10.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:10 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:10.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:10 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:11.639 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:11 vm08 ceph-mon[54477]: Saving service rgw.myrealm1.myzone1 spec with placement count:2 2026-03-09T20:39:11.639 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:11 vm08 ceph-mon[54477]: Deploying daemon rgw.myrealm1.myzone1.vm04.knyogr on vm04 2026-03-09T20:39:11.639 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:11 vm08 ceph-mon[54477]: pgmap v19: 129 pgs: 129 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 255 B/s rd, 511 B/s wr, 1 op/s 2026-03-09T20:39:11.639 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:11 vm08 ceph-mon[54477]: from='client.14634 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:39:11.639 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:11 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:11.639 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:11 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:11.639 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:11 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:11.639 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:11 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm08.paighk", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-09T20:39:11.639 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:11 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm08.paighk", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-09T20:39:11.639 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:11 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:11.639 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:11 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:11.639 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:11 vm08 ceph-mon[54477]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-09T20:39:11.639 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:11 vm08 ceph-mon[54477]: Cluster is now healthy 2026-03-09T20:39:11.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:11 vm03 ceph-mon[52226]: Saving service rgw.myrealm1.myzone1 spec with placement count:2 2026-03-09T20:39:11.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:11 vm03 ceph-mon[52226]: Deploying daemon rgw.myrealm1.myzone1.vm04.knyogr on vm04 2026-03-09T20:39:11.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:11 vm03 ceph-mon[52226]: pgmap v19: 129 pgs: 129 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 255 B/s rd, 511 B/s wr, 1 op/s 2026-03-09T20:39:11.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:11 vm03 ceph-mon[52226]: from='client.14634 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:39:11.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:11 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:11.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:11 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:11.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:11 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:11.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:11 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm08.paighk", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-09T20:39:11.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:11 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm08.paighk", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-09T20:39:11.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:11 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:11.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:11 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:11.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:11 vm03 ceph-mon[52226]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-09T20:39:11.894 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:11 vm03 ceph-mon[52226]: Cluster is now healthy 2026-03-09T20:39:12.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:11 vm04 ceph-mon[53250]: Saving service rgw.myrealm1.myzone1 spec with placement count:2 2026-03-09T20:39:12.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:11 vm04 ceph-mon[53250]: Deploying daemon rgw.myrealm1.myzone1.vm04.knyogr on vm04 2026-03-09T20:39:12.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:11 vm04 ceph-mon[53250]: pgmap v19: 129 pgs: 129 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 255 B/s rd, 511 B/s wr, 1 op/s 2026-03-09T20:39:12.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:11 vm04 ceph-mon[53250]: from='client.14634 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:39:12.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:11 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:12.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:11 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:12.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:11 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:12.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:11 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm08.paighk", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-09T20:39:12.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:11 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.myrealm1.myzone1.vm08.paighk", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-09T20:39:12.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:11 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:12.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:11 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:12.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:11 vm04 ceph-mon[53250]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-09T20:39:12.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:11 vm04 ceph-mon[53250]: Cluster is now healthy 2026-03-09T20:39:13.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:12 vm04 ceph-mon[53250]: Deploying daemon rgw.myrealm1.myzone1.vm08.paighk on vm08 2026-03-09T20:39:13.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:12 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:12 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:12 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:12 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "rgw zone modify", "realm_name": "myrealm1", "zonegroup_name": "myzonegroup1", "zone_name": "myzone1", "realm_token": "ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjg5ODQxMzdjLWViYjktNDYxZC1iMjgyLThkYjVhMjc5MGJmYyIsCiAgICAiZW5kcG9pbnQiOiBudWxsLAogICAgImFjY2Vzc19rZXkiOiAiTjJJSFVLMDAzNkQxMEs5UEQyRVoiLAogICAgInNlY3JldCI6ICJ2aUNRTWlCeVA3NTdyNXJwM1hsenhTd2Rnd1NSWlFSQ2FtZVRud3lNIgp9", "zone_endpoints": ["http://192.168.123.104:5500", "http://192.168.123.108:5500"]}]: dispatch 2026-03-09T20:39:13.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:12 vm04 ceph-mon[53250]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "rgw zone modify", "realm_name": "myrealm1", "zonegroup_name": "myzonegroup1", "zone_name": "myzone1", "realm_token": "ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjg5ODQxMzdjLWViYjktNDYxZC1iMjgyLThkYjVhMjc5MGJmYyIsCiAgICAiZW5kcG9pbnQiOiBudWxsLAogICAgImFjY2Vzc19rZXkiOiAiTjJJSFVLMDAzNkQxMEs5UEQyRVoiLAogICAgInNlY3JldCI6ICJ2aUNRTWlCeVA3NTdyNXJwM1hsenhTd2Rnd1NSWlFSQ2FtZVRud3lNIgp9", "zone_endpoints": ["http://192.168.123.104:5500", "http://192.168.123.108:5500"]}]: dispatch 2026-03-09T20:39:13.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:12 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:12 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:12 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:13.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:12 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:13.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:12 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:13.047 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:12 vm08 ceph-mon[54477]: Deploying daemon rgw.myrealm1.myzone1.vm08.paighk on vm08 2026-03-09T20:39:13.047 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:12 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.047 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:12 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.047 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:12 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.047 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:12 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "rgw zone modify", "realm_name": "myrealm1", "zonegroup_name": "myzonegroup1", "zone_name": "myzone1", "realm_token": "ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjg5ODQxMzdjLWViYjktNDYxZC1iMjgyLThkYjVhMjc5MGJmYyIsCiAgICAiZW5kcG9pbnQiOiBudWxsLAogICAgImFjY2Vzc19rZXkiOiAiTjJJSFVLMDAzNkQxMEs5UEQyRVoiLAogICAgInNlY3JldCI6ICJ2aUNRTWlCeVA3NTdyNXJwM1hsenhTd2Rnd1NSWlFSQ2FtZVRud3lNIgp9", "zone_endpoints": ["http://192.168.123.104:5500", "http://192.168.123.108:5500"]}]: dispatch 2026-03-09T20:39:13.047 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:12 vm08 ceph-mon[54477]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "rgw zone modify", "realm_name": "myrealm1", "zonegroup_name": "myzonegroup1", "zone_name": "myzone1", "realm_token": "ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjg5ODQxMzdjLWViYjktNDYxZC1iMjgyLThkYjVhMjc5MGJmYyIsCiAgICAiZW5kcG9pbnQiOiBudWxsLAogICAgImFjY2Vzc19rZXkiOiAiTjJJSFVLMDAzNkQxMEs5UEQyRVoiLAogICAgInNlY3JldCI6ICJ2aUNRTWlCeVA3NTdyNXJwM1hsenhTd2Rnd1NSWlFSQ2FtZVRud3lNIgp9", "zone_endpoints": ["http://192.168.123.104:5500", "http://192.168.123.108:5500"]}]: dispatch 2026-03-09T20:39:13.047 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:12 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.047 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:12 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.047 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:12 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:13.047 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:12 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:13.047 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:12 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:13.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:12 vm03 ceph-mon[52226]: Deploying daemon rgw.myrealm1.myzone1.vm08.paighk on vm08 2026-03-09T20:39:13.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:12 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:12 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:12 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:12 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "rgw zone modify", "realm_name": "myrealm1", "zonegroup_name": "myzonegroup1", "zone_name": "myzone1", "realm_token": "ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjg5ODQxMzdjLWViYjktNDYxZC1iMjgyLThkYjVhMjc5MGJmYyIsCiAgICAiZW5kcG9pbnQiOiBudWxsLAogICAgImFjY2Vzc19rZXkiOiAiTjJJSFVLMDAzNkQxMEs5UEQyRVoiLAogICAgInNlY3JldCI6ICJ2aUNRTWlCeVA3NTdyNXJwM1hsenhTd2Rnd1NSWlFSQ2FtZVRud3lNIgp9", "zone_endpoints": ["http://192.168.123.104:5500", "http://192.168.123.108:5500"]}]: dispatch 2026-03-09T20:39:13.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:12 vm03 ceph-mon[52226]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "rgw zone modify", "realm_name": "myrealm1", "zonegroup_name": "myzonegroup1", "zone_name": "myzone1", "realm_token": "ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjg5ODQxMzdjLWViYjktNDYxZC1iMjgyLThkYjVhMjc5MGJmYyIsCiAgICAiZW5kcG9pbnQiOiBudWxsLAogICAgImFjY2Vzc19rZXkiOiAiTjJJSFVLMDAzNkQxMEs5UEQyRVoiLAogICAgInNlY3JldCI6ICJ2aUNRTWlCeVA3NTdyNXJwM1hsenhTd2Rnd1NSWlFSQ2FtZVRud3lNIgp9", "zone_endpoints": ["http://192.168.123.104:5500", "http://192.168.123.108:5500"]}]: dispatch 2026-03-09T20:39:13.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:12 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:12 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:12 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:13.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:12 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:13.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:12 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:13.806 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:13 vm08 ceph-mon[54477]: pgmap v20: 129 pgs: 129 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 182 B/s rd, 365 B/s wr, 1 op/s 2026-03-09T20:39:13.806 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:13 vm08 ceph-mon[54477]: Saving service rgw.myrealm1.myzone1 spec with placement count:2 2026-03-09T20:39:13.806 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:13 vm08 ceph-mon[54477]: Checking dashboard <-> RGW credentials 2026-03-09T20:39:13.806 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:13 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.806 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:13 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.806 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:13 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.806 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:13 vm08 ceph-mon[54477]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:39:13.806 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:13 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:13.806 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:13 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:13.806 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:13 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:13.806 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:13 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.806 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:13 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.849 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:13 vm04 ceph-mon[53250]: pgmap v20: 129 pgs: 129 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 182 B/s rd, 365 B/s wr, 1 op/s 2026-03-09T20:39:13.849 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:13 vm04 ceph-mon[53250]: Saving service rgw.myrealm1.myzone1 spec with placement count:2 2026-03-09T20:39:13.849 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:13 vm04 ceph-mon[53250]: Checking dashboard <-> RGW credentials 2026-03-09T20:39:13.849 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:13 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.850 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:13 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.850 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:13 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.850 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:13 vm04 ceph-mon[53250]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:39:13.850 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:13 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:13.850 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:13 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:13.850 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:13 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:13.850 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:13 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:13.850 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:13 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:14.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:13 vm03 ceph-mon[52226]: pgmap v20: 129 pgs: 129 active+clean; 451 KiB data, 81 MiB used, 60 GiB / 60 GiB avail; 182 B/s rd, 365 B/s wr, 1 op/s 2026-03-09T20:39:14.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:13 vm03 ceph-mon[52226]: Saving service rgw.myrealm1.myzone1 spec with placement count:2 2026-03-09T20:39:14.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:13 vm03 ceph-mon[52226]: Checking dashboard <-> RGW credentials 2026-03-09T20:39:14.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:13 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:14.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:13 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:14.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:13 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:14.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:13 vm03 ceph-mon[52226]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:39:14.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:13 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:14.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:13 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:14.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:13 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:14.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:13 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:14.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:13 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:14 vm04 ceph-mon[53250]: Checking dashboard <-> RGW credentials 2026-03-09T20:39:15.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:14 vm04 ceph-mon[53250]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:39:15.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:14 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:14 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:14 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:14 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:14 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:14 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:14 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:14 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:15.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:14 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:15.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:14 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:15.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:14 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:14 vm04 ceph-mon[53250]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:14 vm08 ceph-mon[54477]: Checking dashboard <-> RGW credentials 2026-03-09T20:39:15.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:14 vm08 ceph-mon[54477]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:39:15.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:14 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:14 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:14 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:14 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:14 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:14 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:14 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:14 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:15.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:14 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:15.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:14 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:15.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:14 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.068 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:14 vm08 ceph-mon[54477]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:14 vm03 ceph-mon[52226]: Checking dashboard <-> RGW credentials 2026-03-09T20:39:15.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:14 vm03 ceph-mon[52226]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-09T20:39:15.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:14 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:14 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:14 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:14 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:14 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:14 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:14 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:14 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T20:39:15.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:14 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T20:39:15.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:14 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T20:39:15.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:14 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:14 vm03 ceph-mon[52226]: from='mgr.14421 192.168.123.103:0/3332905493' entity='mgr.a' 2026-03-09T20:39:15.751 INFO:teuthology.orchestra.run.vm03.stderr:+ true 2026-03-09T20:39:15.752 INFO:teuthology.orchestra.run.vm03.stderr:++ ceph rgw realm tokens 2026-03-09T20:39:15.752 INFO:teuthology.orchestra.run.vm03.stderr:++ jq -r '.[0].token' 2026-03-09T20:39:16.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:15 vm04 ceph-mon[53250]: pgmap v21: 129 pgs: 129 active+clean; 460 KiB data, 94 MiB used, 60 GiB / 60 GiB avail; 207 KiB/s rd, 11 KiB/s wr, 363 op/s 2026-03-09T20:39:16.067 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:15 vm08 ceph-mon[54477]: pgmap v21: 129 pgs: 129 active+clean; 460 KiB data, 94 MiB used, 60 GiB / 60 GiB avail; 207 KiB/s rd, 11 KiB/s wr, 363 op/s 2026-03-09T20:39:16.144 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:15 vm03 ceph-mon[52226]: pgmap v21: 129 pgs: 129 active+clean; 460 KiB data, 94 MiB used, 60 GiB / 60 GiB avail; 207 KiB/s rd, 11 KiB/s wr, 363 op/s 2026-03-09T20:39:16.305 INFO:teuthology.orchestra.run.vm03.stdout:ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjg5ODQxMzdjLWViYjktNDYxZC1iMjgyLThkYjVhMjc5MGJmYyIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwNDo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIk4ySUhVSzAwMzZEMTBLOVBEMkVaIiwKICAgICJzZWNyZXQiOiAidmlDUU1pQnlQNzU3cjVycDNYbHp4U3dkZ3dTUlpRUkNhbWVUbnd5TSIKfQ== 2026-03-09T20:39:16.305 INFO:teuthology.orchestra.run.vm03.stderr:+ TOKEN=ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjg5ODQxMzdjLWViYjktNDYxZC1iMjgyLThkYjVhMjc5MGJmYyIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwNDo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIk4ySUhVSzAwMzZEMTBLOVBEMkVaIiwKICAgICJzZWNyZXQiOiAidmlDUU1pQnlQNzU3cjVycDNYbHp4U3dkZ3dTUlpRUkNhbWVUbnd5TSIKfQ== 2026-03-09T20:39:16.305 INFO:teuthology.orchestra.run.vm03.stderr:+ echo ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjg5ODQxMzdjLWViYjktNDYxZC1iMjgyLThkYjVhMjc5MGJmYyIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwNDo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIk4ySUhVSzAwMzZEMTBLOVBEMkVaIiwKICAgICJzZWNyZXQiOiAidmlDUU1pQnlQNzU3cjVycDNYbHp4U3dkZ3dTUlpRUkNhbWVUbnd5TSIKfQ== 2026-03-09T20:39:16.305 INFO:teuthology.orchestra.run.vm03.stderr:+ '[' ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjg5ODQxMzdjLWViYjktNDYxZC1iMjgyLThkYjVhMjc5MGJmYyIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwNDo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIk4ySUhVSzAwMzZEMTBLOVBEMkVaIiwKICAgICJzZWNyZXQiOiAidmlDUU1pQnlQNzU3cjVycDNYbHp4U3dkZ3dTUlpRUkNhbWVUbnd5TSIKfQ== '!=' 'master zone has no endpoint' ']' 2026-03-09T20:39:16.305 INFO:teuthology.orchestra.run.vm03.stderr:+ break 2026-03-09T20:39:16.305 INFO:teuthology.orchestra.run.vm03.stderr:++ ceph rgw realm tokens 2026-03-09T20:39:16.861 INFO:teuthology.orchestra.run.vm03.stderr:+ TOKENS='[ 2026-03-09T20:39:16.861 INFO:teuthology.orchestra.run.vm03.stderr: { 2026-03-09T20:39:16.861 INFO:teuthology.orchestra.run.vm03.stderr: "realm": "myrealm1", 2026-03-09T20:39:16.861 INFO:teuthology.orchestra.run.vm03.stderr: "token": "ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjg5ODQxMzdjLWViYjktNDYxZC1iMjgyLThkYjVhMjc5MGJmYyIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwNDo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIk4ySUhVSzAwMzZEMTBLOVBEMkVaIiwKICAgICJzZWNyZXQiOiAidmlDUU1pQnlQNzU3cjVycDNYbHp4U3dkZ3dTUlpRUkNhbWVUbnd5TSIKfQ==" 2026-03-09T20:39:16.861 INFO:teuthology.orchestra.run.vm03.stderr: } 2026-03-09T20:39:16.861 INFO:teuthology.orchestra.run.vm03.stderr:]' 2026-03-09T20:39:16.861 INFO:teuthology.orchestra.run.vm03.stderr:+ jq --exit-status '.[0].realm == "myrealm1"' 2026-03-09T20:39:16.861 INFO:teuthology.orchestra.run.vm03.stderr:+ echo '[' '{' '"realm":' '"myrealm1",' '"token":' '"ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjg5ODQxMzdjLWViYjktNDYxZC1iMjgyLThkYjVhMjc5MGJmYyIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwNDo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIk4ySUhVSzAwMzZEMTBLOVBEMkVaIiwKICAgICJzZWNyZXQiOiAidmlDUU1pQnlQNzU3cjVycDNYbHp4U3dkZ3dTUlpRUkNhbWVUbnd5TSIKfQ=="' '}' ']' 2026-03-09T20:39:16.862 INFO:teuthology.orchestra.run.vm03.stdout:true 2026-03-09T20:39:16.863 INFO:teuthology.orchestra.run.vm03.stderr:+ echo '[' '{' '"realm":' '"myrealm1",' '"token":' '"ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjg5ODQxMzdjLWViYjktNDYxZC1iMjgyLThkYjVhMjc5MGJmYyIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwNDo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIk4ySUhVSzAwMzZEMTBLOVBEMkVaIiwKICAgICJzZWNyZXQiOiAidmlDUU1pQnlQNzU3cjVycDNYbHp4U3dkZ3dTUlpRUkNhbWVUbnd5TSIKfQ=="' '}' ']' 2026-03-09T20:39:16.863 INFO:teuthology.orchestra.run.vm03.stderr:+ jq --exit-status '.[0].token' 2026-03-09T20:39:16.864 INFO:teuthology.orchestra.run.vm03.stdout:"ewogICAgInJlYWxtX25hbWUiOiAibXlyZWFsbTEiLAogICAgInJlYWxtX2lkIjogIjg5ODQxMzdjLWViYjktNDYxZC1iMjgyLThkYjVhMjc5MGJmYyIsCiAgICAiZW5kcG9pbnQiOiAiaHR0cDovLzE5Mi4xNjguMTIzLjEwNDo1NTAwIiwKICAgICJhY2Nlc3Nfa2V5IjogIk4ySUhVSzAwMzZEMTBLOVBEMkVaIiwKICAgICJzZWNyZXQiOiAidmlDUU1pQnlQNzU3cjVycDNYbHp4U3dkZ3dTUlpRUkNhbWVUbnd5TSIKfQ==" 2026-03-09T20:39:16.865 INFO:teuthology.orchestra.run.vm03.stderr:++ ceph rgw realm tokens 2026-03-09T20:39:16.865 INFO:teuthology.orchestra.run.vm03.stderr:++ jq -r '.[0].token' 2026-03-09T20:39:16.869 INFO:teuthology.orchestra.run.vm03.stderr:++ base64 --decode 2026-03-09T20:39:17.414 INFO:teuthology.orchestra.run.vm03.stderr:+ TOKEN_JSON='{ 2026-03-09T20:39:17.414 INFO:teuthology.orchestra.run.vm03.stderr: "realm_name": "myrealm1", 2026-03-09T20:39:17.414 INFO:teuthology.orchestra.run.vm03.stderr: "realm_id": "8984137c-ebb9-461d-b282-8db5a2790bfc", 2026-03-09T20:39:17.414 INFO:teuthology.orchestra.run.vm03.stderr: "endpoint": "http://192.168.123.104:5500", 2026-03-09T20:39:17.414 INFO:teuthology.orchestra.run.vm03.stderr: "access_key": "N2IHUK0036D10K9PD2EZ", 2026-03-09T20:39:17.414 INFO:teuthology.orchestra.run.vm03.stderr: "secret": "viCQMiByP757r5rp3XlzxSwdgwSRZQRCameTnwyM" 2026-03-09T20:39:17.414 INFO:teuthology.orchestra.run.vm03.stderr:}' 2026-03-09T20:39:17.414 INFO:teuthology.orchestra.run.vm03.stderr:+ jq --exit-status '.realm_name == "myrealm1"' 2026-03-09T20:39:17.414 INFO:teuthology.orchestra.run.vm03.stderr:+ echo '{' '"realm_name":' '"myrealm1",' '"realm_id":' '"8984137c-ebb9-461d-b282-8db5a2790bfc",' '"endpoint":' '"http://192.168.123.104:5500",' '"access_key":' '"N2IHUK0036D10K9PD2EZ",' '"secret":' '"viCQMiByP757r5rp3XlzxSwdgwSRZQRCameTnwyM"' '}' 2026-03-09T20:39:17.416 INFO:teuthology.orchestra.run.vm03.stdout:true 2026-03-09T20:39:17.416 INFO:teuthology.orchestra.run.vm03.stderr:+ echo '{' '"realm_name":' '"myrealm1",' '"realm_id":' '"8984137c-ebb9-461d-b282-8db5a2790bfc",' '"endpoint":' '"http://192.168.123.104:5500",' '"access_key":' '"N2IHUK0036D10K9PD2EZ",' '"secret":' '"viCQMiByP757r5rp3XlzxSwdgwSRZQRCameTnwyM"' '}' 2026-03-09T20:39:17.416 INFO:teuthology.orchestra.run.vm03.stderr:+ jq --exit-status '.endpoint | test("http://.+:\\d+")' 2026-03-09T20:39:17.418 INFO:teuthology.orchestra.run.vm03.stdout:true 2026-03-09T20:39:17.418 INFO:teuthology.orchestra.run.vm03.stderr:+ echo '{' '"realm_name":' '"myrealm1",' '"realm_id":' '"8984137c-ebb9-461d-b282-8db5a2790bfc",' '"endpoint":' '"http://192.168.123.104:5500",' '"access_key":' '"N2IHUK0036D10K9PD2EZ",' '"secret":' '"viCQMiByP757r5rp3XlzxSwdgwSRZQRCameTnwyM"' '}' 2026-03-09T20:39:17.418 INFO:teuthology.orchestra.run.vm03.stderr:+ jq --exit-status '.realm_id | test("^[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}$")' 2026-03-09T20:39:17.420 INFO:teuthology.orchestra.run.vm03.stdout:true 2026-03-09T20:39:17.420 INFO:teuthology.orchestra.run.vm03.stderr:+ echo '{' '"realm_name":' '"myrealm1",' '"realm_id":' '"8984137c-ebb9-461d-b282-8db5a2790bfc",' '"endpoint":' '"http://192.168.123.104:5500",' '"access_key":' '"N2IHUK0036D10K9PD2EZ",' '"secret":' '"viCQMiByP757r5rp3XlzxSwdgwSRZQRCameTnwyM"' '}' 2026-03-09T20:39:17.420 INFO:teuthology.orchestra.run.vm03.stderr:+ jq --exit-status .access_key 2026-03-09T20:39:17.422 INFO:teuthology.orchestra.run.vm03.stdout:"N2IHUK0036D10K9PD2EZ" 2026-03-09T20:39:17.422 INFO:teuthology.orchestra.run.vm03.stderr:+ echo '{' '"realm_name":' '"myrealm1",' '"realm_id":' '"8984137c-ebb9-461d-b282-8db5a2790bfc",' '"endpoint":' '"http://192.168.123.104:5500",' '"access_key":' '"N2IHUK0036D10K9PD2EZ",' '"secret":' '"viCQMiByP757r5rp3XlzxSwdgwSRZQRCameTnwyM"' '}' 2026-03-09T20:39:17.422 INFO:teuthology.orchestra.run.vm03.stderr:+ jq --exit-status .secret 2026-03-09T20:39:17.424 INFO:teuthology.orchestra.run.vm03.stdout:"viCQMiByP757r5rp3XlzxSwdgwSRZQRCameTnwyM" 2026-03-09T20:39:17.478 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-09T20:39:17.480 INFO:tasks.cephadm:Teardown begin 2026-03-09T20:39:17.480 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:39:17.506 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:39:17.532 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:39:17.559 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-09T20:39:17.559 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd -- ceph mgr module disable cephadm 2026-03-09T20:39:17.716 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:17 vm03 ceph-mon[52226]: from='client.14952 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:39:17.716 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:17 vm03 ceph-mon[52226]: pgmap v22: 129 pgs: 129 active+clean; 460 KiB data, 94 MiB used, 60 GiB / 60 GiB avail; 171 KiB/s rd, 9.2 KiB/s wr, 299 op/s 2026-03-09T20:39:17.716 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:17 vm03 ceph-mon[52226]: from='client.15054 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:39:17.725 INFO:teuthology.orchestra.run.vm03.stderr:Inferring config /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/mon.a/config 2026-03-09T20:39:17.741 INFO:teuthology.orchestra.run.vm03.stderr:Error: statfs /etc/ceph/ceph.client.admin.keyring: no such file or directory 2026-03-09T20:39:17.760 DEBUG:teuthology.orchestra.run:got remote process result: 125 2026-03-09T20:39:17.760 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-09T20:39:17.760 DEBUG:teuthology.orchestra.run.vm03:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T20:39:17.777 DEBUG:teuthology.orchestra.run.vm04:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T20:39:17.792 DEBUG:teuthology.orchestra.run.vm08:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T20:39:17.810 INFO:tasks.cephadm:Stopping all daemons... 2026-03-09T20:39:17.810 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-09T20:39:17.810 DEBUG:teuthology.orchestra.run.vm03:> sudo systemctl stop ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mon.a 2026-03-09T20:39:17.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:17 vm08 ceph-mon[54477]: from='client.14952 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:39:17.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:17 vm08 ceph-mon[54477]: pgmap v22: 129 pgs: 129 active+clean; 460 KiB data, 94 MiB used, 60 GiB / 60 GiB avail; 171 KiB/s rd, 9.2 KiB/s wr, 299 op/s 2026-03-09T20:39:17.817 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:17 vm08 ceph-mon[54477]: from='client.15054 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:39:18.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:17 vm04 ceph-mon[53250]: from='client.14952 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:39:18.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:17 vm04 ceph-mon[53250]: pgmap v22: 129 pgs: 129 active+clean; 460 KiB data, 94 MiB used, 60 GiB / 60 GiB avail; 171 KiB/s rd, 9.2 KiB/s wr, 299 op/s 2026-03-09T20:39:18.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:17 vm04 ceph-mon[53250]: from='client.15054 -' entity='client.admin' cmd=[{"prefix": "rgw realm tokens", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T20:39:18.095 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:17 vm03 systemd[1]: Stopping Ceph mon.a for 8176208c-1bf7-11f1-86ba-4f626c793dbd... 2026-03-09T20:39:18.095 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:17 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-a[52222]: 2026-03-09T20:39:17.928+0000 7f8ca0a1a640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T20:39:18.095 INFO:journalctl@ceph.mon.a.vm03.stdout:Mar 09 20:39:17 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-a[52222]: 2026-03-09T20:39:17.928+0000 7f8ca0a1a640 -1 mon.a@0(leader) e3 *** Got Signal Terminated *** 2026-03-09T20:39:18.184 DEBUG:teuthology.orchestra.run.vm03:> sudo pkill -f 'journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mon.a.service' 2026-03-09T20:39:18.221 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:39:18.221 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-09T20:39:18.221 INFO:tasks.cephadm.mon.c:Stopping mon.b... 2026-03-09T20:39:18.221 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mon.b 2026-03-09T20:39:18.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:18 vm04 systemd[1]: Stopping Ceph mon.b for 8176208c-1bf7-11f1-86ba-4f626c793dbd... 2026-03-09T20:39:18.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:18 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-b[53246]: 2026-03-09T20:39:18.331+0000 7ffa6c4d6640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T20:39:18.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:18 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-b[53246]: 2026-03-09T20:39:18.331+0000 7ffa6c4d6640 -1 mon.b@1(peon) e3 *** Got Signal Terminated *** 2026-03-09T20:39:18.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:18 vm04 podman[61895]: 2026-03-09 20:39:18.443819129 +0000 UTC m=+0.125926094 container died b3a4fbf46922a3bb5a2cd5906aeca73da1429ecad41b4a4fd2106d62744cd9bd (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-b, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, OSD_FLAVOR=default, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-09T20:39:18.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:18 vm04 podman[61895]: 2026-03-09 20:39:18.462590114 +0000 UTC m=+0.144697079 container remove b3a4fbf46922a3bb5a2cd5906aeca73da1429ecad41b4a4fd2106d62744cd9bd (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-b, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) 2026-03-09T20:39:18.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 20:39:18 vm04 bash[61895]: ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-b 2026-03-09T20:39:18.531 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mon.b.service' 2026-03-09T20:39:18.568 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:39:18.568 INFO:tasks.cephadm.mon.c:Stopped mon.b 2026-03-09T20:39:18.568 INFO:tasks.cephadm.mon.c:Stopping mon.c... 2026-03-09T20:39:18.568 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mon.c 2026-03-09T20:39:18.917 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:18 vm08 systemd[1]: Stopping Ceph mon.c for 8176208c-1bf7-11f1-86ba-4f626c793dbd... 2026-03-09T20:39:18.917 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:18 vm08 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-c[54467]: 2026-03-09T20:39:18.669+0000 7fcf4ca35640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.c -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T20:39:18.917 INFO:journalctl@ceph.mon.c.vm08.stdout:Mar 09 20:39:18 vm08 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mon-c[54467]: 2026-03-09T20:39:18.669+0000 7fcf4ca35640 -1 mon.c@2(peon) e3 *** Got Signal Terminated *** 2026-03-09T20:39:18.996 DEBUG:teuthology.orchestra.run.vm08:> sudo pkill -f 'journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mon.c.service' 2026-03-09T20:39:19.039 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:39:19.039 INFO:tasks.cephadm.mon.c:Stopped mon.c 2026-03-09T20:39:19.039 INFO:tasks.cephadm.mgr.a:Stopping mgr.a... 2026-03-09T20:39:19.040 DEBUG:teuthology.orchestra.run.vm03:> sudo systemctl stop ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mgr.a 2026-03-09T20:39:19.291 DEBUG:teuthology.orchestra.run.vm03:> sudo pkill -f 'journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mgr.a.service' 2026-03-09T20:39:19.314 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:39:19 vm03 systemd[1]: Stopping Ceph mgr.a for 8176208c-1bf7-11f1-86ba-4f626c793dbd... 2026-03-09T20:39:19.314 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:39:19 vm03 podman[69320]: 2026-03-09 20:39:19.182152682 +0000 UTC m=+0.055977184 container died c72f3f9f23987ee38ddac06c6fad7ec61bfb8eecc61d004eabb2d56ca11746c4 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, CEPH_REF=squid, OSD_FLAVOR=default) 2026-03-09T20:39:19.314 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:39:19 vm03 podman[69320]: 2026-03-09 20:39:19.214770593 +0000 UTC m=+0.088595096 container remove c72f3f9f23987ee38ddac06c6fad7ec61bfb8eecc61d004eabb2d56ca11746c4 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, ceph=True, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T20:39:19.314 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:39:19 vm03 bash[69320]: ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-mgr-a 2026-03-09T20:39:19.314 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:39:19 vm03 systemd[1]: ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mgr.a.service: Deactivated successfully. 2026-03-09T20:39:19.314 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:39:19 vm03 systemd[1]: Stopped Ceph mgr.a for 8176208c-1bf7-11f1-86ba-4f626c793dbd. 2026-03-09T20:39:19.314 INFO:journalctl@ceph.mgr.a.vm03.stdout:Mar 09 20:39:19 vm03 systemd[1]: ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mgr.a.service: Consumed 37.975s CPU time. 2026-03-09T20:39:19.321 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:39:19.322 INFO:tasks.cephadm.mgr.a:Stopped mgr.a 2026-03-09T20:39:19.322 INFO:tasks.cephadm.mgr.b:Stopping mgr.b... 2026-03-09T20:39:19.322 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mgr.b 2026-03-09T20:39:19.468 INFO:journalctl@ceph.mgr.b.vm04.stdout:Mar 09 20:39:19 vm04 systemd[1]: Stopping Ceph mgr.b for 8176208c-1bf7-11f1-86ba-4f626c793dbd... 2026-03-09T20:39:19.679 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@mgr.b.service' 2026-03-09T20:39:19.723 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:39:19.723 INFO:tasks.cephadm.mgr.b:Stopped mgr.b 2026-03-09T20:39:19.723 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-09T20:39:19.723 DEBUG:teuthology.orchestra.run.vm03:> sudo systemctl stop ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@osd.0 2026-03-09T20:39:20.144 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:39:19 vm03 systemd[1]: Stopping Ceph osd.0 for 8176208c-1bf7-11f1-86ba-4f626c793dbd... 2026-03-09T20:39:20.144 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:39:19 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-0[60583]: 2026-03-09T20:39:19.826+0000 7f3d6a41f640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T20:39:20.144 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:39:19 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-0[60583]: 2026-03-09T20:39:19.826+0000 7f3d6a41f640 -1 osd.0 32 *** Got signal Terminated *** 2026-03-09T20:39:20.144 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:39:19 vm03 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-0[60583]: 2026-03-09T20:39:19.826+0000 7f3d6a41f640 -1 osd.0 32 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T20:39:25.114 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:39:24 vm03 podman[69425]: 2026-03-09 20:39:24.862624655 +0000 UTC m=+5.049309113 container died cb46228fd016dd10a9e63996346b313f01d2a11c0f59a71acd3ba48ae05e38ca (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-0, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, OSD_FLAVOR=default) 2026-03-09T20:39:25.114 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:39:24 vm03 podman[69425]: 2026-03-09 20:39:24.885119892 +0000 UTC m=+5.071804350 container remove cb46228fd016dd10a9e63996346b313f01d2a11c0f59a71acd3ba48ae05e38ca (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-09T20:39:25.114 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:39:24 vm03 bash[69425]: ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-0 2026-03-09T20:39:25.114 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:39:25 vm03 podman[69762]: 2026-03-09 20:39:25.021223042 +0000 UTC m=+0.017272825 container create 3c879270f7e679778edba23e64fdde0c2fe3a94cb9bb5ae0c3d386b85fb53095 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-0-deactivate, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, CEPH_REF=squid, ceph=True, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-09T20:39:25.114 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:39:25 vm03 podman[69762]: 2026-03-09 20:39:25.074703604 +0000 UTC m=+0.070753387 container init 3c879270f7e679778edba23e64fdde0c2fe3a94cb9bb5ae0c3d386b85fb53095 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-0-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-09T20:39:25.114 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:39:25 vm03 podman[69762]: 2026-03-09 20:39:25.078332865 +0000 UTC m=+0.074382648 container start 3c879270f7e679778edba23e64fdde0c2fe3a94cb9bb5ae0c3d386b85fb53095 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-0-deactivate, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, ceph=True, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, OSD_FLAVOR=default, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223) 2026-03-09T20:39:25.114 INFO:journalctl@ceph.osd.0.vm03.stdout:Mar 09 20:39:25 vm03 podman[69762]: 2026-03-09 20:39:25.085628598 +0000 UTC m=+0.081678381 container attach 3c879270f7e679778edba23e64fdde0c2fe3a94cb9bb5ae0c3d386b85fb53095 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-0-deactivate, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2) 2026-03-09T20:39:25.236 DEBUG:teuthology.orchestra.run.vm03:> sudo pkill -f 'journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@osd.0.service' 2026-03-09T20:39:25.268 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:39:25.268 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-09T20:39:25.268 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-09T20:39:25.268 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@osd.1 2026-03-09T20:39:25.755 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:39:25 vm04 systemd[1]: Stopping Ceph osd.1 for 8176208c-1bf7-11f1-86ba-4f626c793dbd... 2026-03-09T20:39:25.755 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:39:25 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-1[57533]: 2026-03-09T20:39:25.370+0000 7fc033bd1640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T20:39:25.755 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:39:25 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-1[57533]: 2026-03-09T20:39:25.370+0000 7fc033bd1640 -1 osd.1 32 *** Got signal Terminated *** 2026-03-09T20:39:25.755 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:39:25 vm04 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-1[57533]: 2026-03-09T20:39:25.370+0000 7fc033bd1640 -1 osd.1 32 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T20:39:30.656 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:39:30 vm04 podman[62102]: 2026-03-09 20:39:30.394847485 +0000 UTC m=+5.041331136 container died 128e6408efd0967e12dd1dc51c64152f2c4d84fb488d3a01428d10be157fa5c2 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-1, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-09T20:39:30.656 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:39:30 vm04 podman[62102]: 2026-03-09 20:39:30.424070974 +0000 UTC m=+5.070554625 container remove 128e6408efd0967e12dd1dc51c64152f2c4d84fb488d3a01428d10be157fa5c2 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-1, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-09T20:39:30.656 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:39:30 vm04 bash[62102]: ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-1 2026-03-09T20:39:30.656 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:39:30 vm04 podman[62171]: 2026-03-09 20:39:30.562709479 +0000 UTC m=+0.016283860 container create aa9c3b5099639cfd91c7892e40b52c45ba8b5620d2263f85a10acc0163d824c1 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-1-deactivate, ceph=True, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223) 2026-03-09T20:39:30.656 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:39:30 vm04 podman[62171]: 2026-03-09 20:39:30.602917318 +0000 UTC m=+0.056491699 container init aa9c3b5099639cfd91c7892e40b52c45ba8b5620d2263f85a10acc0163d824c1 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-1-deactivate, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-09T20:39:30.656 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:39:30 vm04 podman[62171]: 2026-03-09 20:39:30.606500273 +0000 UTC m=+0.060074654 container start aa9c3b5099639cfd91c7892e40b52c45ba8b5620d2263f85a10acc0163d824c1 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-1-deactivate, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, ceph=True, org.label-schema.build-date=20260223, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-09T20:39:30.656 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 09 20:39:30 vm04 podman[62171]: 2026-03-09 20:39:30.610669217 +0000 UTC m=+0.064243609 container attach aa9c3b5099639cfd91c7892e40b52c45ba8b5620d2263f85a10acc0163d824c1 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-1-deactivate, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T20:39:30.768 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@osd.1.service' 2026-03-09T20:39:30.807 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:39:30.807 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-09T20:39:30.807 INFO:tasks.cephadm.osd.2:Stopping osd.2... 2026-03-09T20:39:30.807 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@osd.2 2026-03-09T20:39:31.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:39:30 vm08 systemd[1]: Stopping Ceph osd.2 for 8176208c-1bf7-11f1-86ba-4f626c793dbd... 2026-03-09T20:39:31.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:39:30 vm08 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-2[58473]: 2026-03-09T20:39:30.947+0000 7f419aa85640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T20:39:31.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:39:30 vm08 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-2[58473]: 2026-03-09T20:39:30.947+0000 7f419aa85640 -1 osd.2 32 *** Got signal Terminated *** 2026-03-09T20:39:31.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:39:30 vm08 ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-2[58473]: 2026-03-09T20:39:30.947+0000 7f419aa85640 -1 osd.2 32 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T20:39:36.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:39:35 vm08 podman[62702]: 2026-03-09 20:39:35.980675627 +0000 UTC m=+5.087837121 container died 69eaf5847494efd1cd6aacd877b4befad1ba1c4929c630c880e5f7642758d606 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, org.label-schema.build-date=20260223) 2026-03-09T20:39:36.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:39:36 vm08 podman[62702]: 2026-03-09 20:39:36.008462385 +0000 UTC m=+5.115623869 container remove 69eaf5847494efd1cd6aacd877b4befad1ba1c4929c630c880e5f7642758d606 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-2, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0) 2026-03-09T20:39:36.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:39:36 vm08 bash[62702]: ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-2 2026-03-09T20:39:36.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:39:36 vm08 podman[63076]: 2026-03-09 20:39:36.135797778 +0000 UTC m=+0.018078652 container create ce243b4fd48d36b65c37de40f51b7867cebcb257b1df127ec24d9a7847bc6ec4 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-2-deactivate, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, CEPH_REF=squid, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-09T20:39:36.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:39:36 vm08 podman[63076]: 2026-03-09 20:39:36.17639636 +0000 UTC m=+0.058677234 container init ce243b4fd48d36b65c37de40f51b7867cebcb257b1df127ec24d9a7847bc6ec4 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-2-deactivate, CEPH_REF=squid, OSD_FLAVOR=default, ceph=True, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-09T20:39:36.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:39:36 vm08 podman[63076]: 2026-03-09 20:39:36.18023878 +0000 UTC m=+0.062519654 container start ce243b4fd48d36b65c37de40f51b7867cebcb257b1df127ec24d9a7847bc6ec4 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-2-deactivate, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-09T20:39:36.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:39:36 vm08 podman[63076]: 2026-03-09 20:39:36.188105351 +0000 UTC m=+0.070386236 container attach ce243b4fd48d36b65c37de40f51b7867cebcb257b1df127ec24d9a7847bc6ec4 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-2-deactivate, CEPH_REF=squid, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS) 2026-03-09T20:39:36.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:39:36 vm08 podman[63076]: 2026-03-09 20:39:36.129046344 +0000 UTC m=+0.011327228 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T20:39:36.318 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 20:39:36 vm08 podman[63076]: 2026-03-09 20:39:36.317807243 +0000 UTC m=+0.200088117 container died ce243b4fd48d36b65c37de40f51b7867cebcb257b1df127ec24d9a7847bc6ec4 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd-osd-2-deactivate, CEPH_REF=squid, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS) 2026-03-09T20:39:36.343 DEBUG:teuthology.orchestra.run.vm08:> sudo pkill -f 'journalctl -f -n 0 -u ceph-8176208c-1bf7-11f1-86ba-4f626c793dbd@osd.2.service' 2026-03-09T20:39:36.380 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T20:39:36.380 INFO:tasks.cephadm.osd.2:Stopped osd.2 2026-03-09T20:39:36.380 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd --force --keep-logs 2026-03-09T20:39:36.507 INFO:teuthology.orchestra.run.vm03.stdout:Deleting cluster with fsid: 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:39:37.629 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd --force --keep-logs 2026-03-09T20:39:37.764 INFO:teuthology.orchestra.run.vm04.stdout:Deleting cluster with fsid: 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:39:49.514 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd --force --keep-logs 2026-03-09T20:39:49.640 INFO:teuthology.orchestra.run.vm08.stdout:Deleting cluster with fsid: 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:40:00.913 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:40:00.938 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:40:00.966 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T20:40:00.991 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-09T20:40:00.992 DEBUG:teuthology.misc:Transferring archived files from vm03:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/646/remote/vm03/crash 2026-03-09T20:40:00.992 DEBUG:teuthology.orchestra.run.vm03:> sudo tar c -f - -C /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/crash -- . 2026-03-09T20:40:01.016 INFO:teuthology.orchestra.run.vm03.stderr:tar: /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/crash: Cannot open: No such file or directory 2026-03-09T20:40:01.016 INFO:teuthology.orchestra.run.vm03.stderr:tar: Error is not recoverable: exiting now 2026-03-09T20:40:01.018 DEBUG:teuthology.misc:Transferring archived files from vm04:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/646/remote/vm04/crash 2026-03-09T20:40:01.018 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/crash -- . 2026-03-09T20:40:01.044 INFO:teuthology.orchestra.run.vm04.stderr:tar: /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/crash: Cannot open: No such file or directory 2026-03-09T20:40:01.044 INFO:teuthology.orchestra.run.vm04.stderr:tar: Error is not recoverable: exiting now 2026-03-09T20:40:01.045 DEBUG:teuthology.misc:Transferring archived files from vm08:/var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/646/remote/vm08/crash 2026-03-09T20:40:01.045 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/crash -- . 2026-03-09T20:40:01.072 INFO:teuthology.orchestra.run.vm08.stderr:tar: /var/lib/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/crash: Cannot open: No such file or directory 2026-03-09T20:40:01.072 INFO:teuthology.orchestra.run.vm08.stderr:tar: Error is not recoverable: exiting now 2026-03-09T20:40:01.073 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-09T20:40:01.074 DEBUG:teuthology.orchestra.run.vm03:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v MON_DOWN | egrep -v 'mons down' | egrep -v 'mon down' | egrep -v 'out of quorum' | egrep -v CEPHADM_STRAY_DAEMON | head -n 1 2026-03-09T20:40:01.103 INFO:tasks.cephadm:Compressing logs... 2026-03-09T20:40:01.103 DEBUG:teuthology.orchestra.run.vm03:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T20:40:01.145 DEBUG:teuthology.orchestra.run.vm04:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T20:40:01.146 DEBUG:teuthology.orchestra.run.vm08:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T20:40:01.170 INFO:teuthology.orchestra.run.vm04.stderr:find: ‘/var/log/rbd-target-api’: No such file or directorygzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T20:40:01.170 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T20:40:01.171 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-volume.log 2026-03-09T20:40:01.172 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mon.b.log 2026-03-09T20:40:01.172 INFO:teuthology.orchestra.run.vm04.stderr: 73.5% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T20:40:01.173 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.audit.log 2026-03-09T20:40:01.173 INFO:teuthology.orchestra.run.vm08.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T20:40:01.174 INFO:teuthology.orchestra.run.vm08.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-09T20:40:01.174 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-volume.log 2026-03-09T20:40:01.175 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mon.b.log: 95.3% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-volume.log.gz 2026-03-09T20:40:01.175 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-volume.log: 74.1% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T20:40:01.175 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mon.c.log 2026-03-09T20:40:01.175 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.log 2026-03-09T20:40:01.176 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T20:40:01.176 INFO:teuthology.orchestra.run.vm03.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-09T20:40:01.176 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.audit.log: 90.2% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.audit.log.gz 2026-03-09T20:40:01.177 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.cephadm.log 2026-03-09T20:40:01.177 INFO:teuthology.orchestra.run.vm08.stderr: 95.3% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-volume.log.gz 2026-03-09T20:40:01.178 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.log 2026-03-09T20:40:01.178 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.audit.log 2026-03-09T20:40:01.178 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.log: 82.9% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.log.gz 2026-03-09T20:40:01.178 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mgr.b.log 2026-03-09T20:40:01.178 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.cephadm.log: 80.9% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.cephadm.log.gz 2026-03-09T20:40:01.178 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mon.c.log: /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.log: 82.3% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.log.gz 2026-03-09T20:40:01.179 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mon.a.log 2026-03-09T20:40:01.179 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.log 2026-03-09T20:40:01.179 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-osd.1.log 2026-03-09T20:40:01.180 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/cephadm.log: /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mon.a.log: 83.2% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T20:40:01.180 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mgr.a.log 2026-03-09T20:40:01.181 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.log: 83.4% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.log.gz 2026-03-09T20:40:01.181 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.audit.log 2026-03-09T20:40:01.182 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.cephadm.log 2026-03-09T20:40:01.183 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.audit.log: 90.0% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.audit.log.gz 2026-03-09T20:40:01.183 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-osd.2.log 2026-03-09T20:40:01.183 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mgr.b.log: gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-client.rgw.myrealm1.myzone1.vm04.knyogr.log 2026-03-09T20:40:01.184 INFO:teuthology.orchestra.run.vm04.stderr: 90.0% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mgr.b.log.gz 2026-03-09T20:40:01.184 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.cephadm.log: 79.6% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.cephadm.log.gz 2026-03-09T20:40:01.184 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-client.rgw.myrealm1.myzone1.vm08.paighk.log 2026-03-09T20:40:01.187 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mgr.a.log: gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.cephadm.log 2026-03-09T20:40:01.190 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.audit.log: 90.1% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.audit.log.gz 2026-03-09T20:40:01.190 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-osd.1.log: /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-client.rgw.myrealm1.myzone1.vm04.knyogr.log: 63.5% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-client.rgw.myrealm1.myzone1.vm04.knyogr.log.gz 2026-03-09T20:40:01.193 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-osd.2.log: /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-client.rgw.myrealm1.myzone1.vm08.paighk.log: 63.3% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-client.rgw.myrealm1.myzone1.vm08.paighk.log.gz 2026-03-09T20:40:01.194 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-volume.log 2026-03-09T20:40:01.196 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.cephadm.log: 82.8% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph.cephadm.log.gz 2026-03-09T20:40:01.199 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-osd.0.log 2026-03-09T20:40:01.205 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-volume.log: 95.1% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-volume.log.gz 2026-03-09T20:40:01.208 INFO:teuthology.orchestra.run.vm08.stderr: 92.4% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mon.c.log.gz 2026-03-09T20:40:01.214 INFO:teuthology.orchestra.run.vm04.stderr: 92.5% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mon.b.log.gz 2026-03-09T20:40:01.271 INFO:teuthology.orchestra.run.vm03.stderr:/var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-osd.0.log: 89.4% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mgr.a.log.gz 2026-03-09T20:40:01.285 INFO:teuthology.orchestra.run.vm03.stderr: 92.0% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-mon.a.log.gz 2026-03-09T20:40:01.374 INFO:teuthology.orchestra.run.vm08.stderr: 94.5% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-osd.2.log.gz 2026-03-09T20:40:01.375 INFO:teuthology.orchestra.run.vm08.stderr: 2026-03-09T20:40:01.375 INFO:teuthology.orchestra.run.vm08.stderr:real 0m0.213s 2026-03-09T20:40:01.375 INFO:teuthology.orchestra.run.vm08.stderr:user 0m0.213s 2026-03-09T20:40:01.375 INFO:teuthology.orchestra.run.vm08.stderr:sys 0m0.032s 2026-03-09T20:40:01.384 INFO:teuthology.orchestra.run.vm04.stderr: 94.4% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-osd.1.log.gz 2026-03-09T20:40:01.386 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T20:40:01.386 INFO:teuthology.orchestra.run.vm04.stderr:real 0m0.226s 2026-03-09T20:40:01.386 INFO:teuthology.orchestra.run.vm04.stderr:user 0m0.246s 2026-03-09T20:40:01.386 INFO:teuthology.orchestra.run.vm04.stderr:sys 0m0.027s 2026-03-09T20:40:01.418 INFO:teuthology.orchestra.run.vm03.stderr: 94.5% -- replaced with /var/log/ceph/8176208c-1bf7-11f1-86ba-4f626c793dbd/ceph-osd.0.log.gz 2026-03-09T20:40:01.420 INFO:teuthology.orchestra.run.vm03.stderr: 2026-03-09T20:40:01.420 INFO:teuthology.orchestra.run.vm03.stderr:real 0m0.257s 2026-03-09T20:40:01.420 INFO:teuthology.orchestra.run.vm03.stderr:user 0m0.330s 2026-03-09T20:40:01.420 INFO:teuthology.orchestra.run.vm03.stderr:sys 0m0.038s 2026-03-09T20:40:01.420 INFO:tasks.cephadm:Archiving logs... 2026-03-09T20:40:01.420 DEBUG:teuthology.misc:Transferring archived files from vm03:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/646/remote/vm03/log 2026-03-09T20:40:01.420 DEBUG:teuthology.orchestra.run.vm03:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T20:40:01.505 DEBUG:teuthology.misc:Transferring archived files from vm04:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/646/remote/vm04/log 2026-03-09T20:40:01.505 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T20:40:01.541 DEBUG:teuthology.misc:Transferring archived files from vm08:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/646/remote/vm08/log 2026-03-09T20:40:01.541 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T20:40:01.579 INFO:tasks.cephadm:Removing cluster... 2026-03-09T20:40:01.579 DEBUG:teuthology.orchestra.run.vm03:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd --force 2026-03-09T20:40:01.707 INFO:teuthology.orchestra.run.vm03.stdout:Deleting cluster with fsid: 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:40:01.930 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd --force 2026-03-09T20:40:02.070 INFO:teuthology.orchestra.run.vm04.stdout:Deleting cluster with fsid: 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:40:02.296 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 8176208c-1bf7-11f1-86ba-4f626c793dbd --force 2026-03-09T20:40:02.424 INFO:teuthology.orchestra.run.vm08.stdout:Deleting cluster with fsid: 8176208c-1bf7-11f1-86ba-4f626c793dbd 2026-03-09T20:40:02.634 INFO:tasks.cephadm:Removing cephadm ... 2026-03-09T20:40:02.634 DEBUG:teuthology.orchestra.run.vm03:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T20:40:02.649 DEBUG:teuthology.orchestra.run.vm04:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T20:40:02.665 DEBUG:teuthology.orchestra.run.vm08:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T20:40:02.679 INFO:tasks.cephadm:Teardown complete 2026-03-09T20:40:02.680 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-09T20:40:02.682 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-09T20:40:02.682 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T20:40:02.691 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T20:40:02.707 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T20:40:02.752 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T20:40:02.753 DEBUG:teuthology.orchestra.run.vm03:> 2026-03-09T20:40:02.753 DEBUG:teuthology.orchestra.run.vm03:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T20:40:02.753 DEBUG:teuthology.orchestra.run.vm03:> sudo yum -y remove $d || true 2026-03-09T20:40:02.753 DEBUG:teuthology.orchestra.run.vm03:> done 2026-03-09T20:40:02.758 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T20:40:02.759 DEBUG:teuthology.orchestra.run.vm04:> 2026-03-09T20:40:02.759 DEBUG:teuthology.orchestra.run.vm04:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T20:40:02.759 DEBUG:teuthology.orchestra.run.vm04:> sudo yum -y remove $d || true 2026-03-09T20:40:02.759 DEBUG:teuthology.orchestra.run.vm04:> done 2026-03-09T20:40:02.764 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T20:40:02.764 DEBUG:teuthology.orchestra.run.vm08:> 2026-03-09T20:40:02.764 DEBUG:teuthology.orchestra.run.vm08:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T20:40:02.764 DEBUG:teuthology.orchestra.run.vm08:> sudo yum -y remove $d || true 2026-03-09T20:40:02.764 DEBUG:teuthology.orchestra.run.vm08:> done 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout:Removing unused dependencies: 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout:Remove 2 Packages 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 39 M 2026-03-09T20:40:02.943 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T20:40:02.945 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T20:40:02.945 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T20:40:02.953 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:02.954 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:02.954 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T20:40:02.954 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:02.954 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T20:40:02.954 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T20:40:02.954 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T20:40:02.954 INFO:teuthology.orchestra.run.vm04.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T20:40:02.954 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:02.954 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T20:40:02.954 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:02.954 INFO:teuthology.orchestra.run.vm04.stdout:Remove 2 Packages 2026-03-09T20:40:02.954 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:02.954 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 39 M 2026-03-09T20:40:02.954 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T20:40:02.956 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T20:40:02.956 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T20:40:02.958 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T20:40:02.958 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T20:40:02.968 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:02.968 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:02.968 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:40:02.968 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:02.969 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:40:02.969 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T20:40:02.969 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-09T20:40:02.969 INFO:teuthology.orchestra.run.vm08.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T20:40:02.969 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:02.969 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:40:02.969 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:02.969 INFO:teuthology.orchestra.run.vm08.stdout:Remove 2 Packages 2026-03-09T20:40:02.969 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:02.969 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 39 M 2026-03-09T20:40:02.969 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T20:40:02.969 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:40:02.969 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T20:40:02.971 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:40:02.971 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:40:02.984 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:40:02.984 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:40:02.988 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T20:40:03.000 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T20:40:03.011 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:03.011 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:03.011 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T20:40:03.011 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T20:40:03.011 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T20:40:03.011 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:03.014 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:40:03.015 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:03.023 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:03.023 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:03.023 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T20:40:03.023 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T20:40:03.023 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T20:40:03.023 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:03.024 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:03.027 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:03.035 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:03.036 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:03.036 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:03.036 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T20:40:03.036 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T20:40:03.036 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T20:40:03.036 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:03.038 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:40:03.039 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:03.048 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:03.050 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:40:03.062 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:40:03.116 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:40:03.116 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:03.123 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:40:03.123 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:03.143 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:40:03.143 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:03.167 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:40:03.167 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:03.167 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T20:40:03.167 INFO:teuthology.orchestra.run.vm03.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T20:40:03.167 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:03.167 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:03.173 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:40:03.173 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:03.174 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T20:40:03.174 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T20:40:03.174 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:03.174 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:03.200 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T20:40:03.200 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:03.200 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:40:03.200 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T20:40:03.200 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:03.200 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:03.382 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout:Removing unused dependencies: 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout:Remove 4 Packages 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 212 M 2026-03-09T20:40:03.383 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout:Remove 4 Packages 2026-03-09T20:40:03.384 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:03.385 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 212 M 2026-03-09T20:40:03.385 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T20:40:03.386 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T20:40:03.386 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T20:40:03.387 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T20:40:03.388 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T20:40:03.404 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:03.404 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout:Remove 4 Packages 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 212 M 2026-03-09T20:40:03.405 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:40:03.408 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:40:03.408 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:40:03.408 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T20:40:03.409 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T20:40:03.410 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T20:40:03.410 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T20:40:03.431 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:40:03.431 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:40:03.473 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T20:40:03.474 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T20:40:03.480 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T20:40:03.481 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T20:40:03.482 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T20:40:03.484 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T20:40:03.485 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T20:40:03.487 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T20:40:03.493 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:40:03.498 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T20:40:03.501 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T20:40:03.501 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T20:40:03.502 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T20:40:03.504 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T20:40:03.521 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T20:40:03.566 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T20:40:03.566 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T20:40:03.566 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T20:40:03.566 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T20:40:03.570 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T20:40:03.570 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T20:40:03.570 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T20:40:03.570 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T20:40:03.598 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T20:40:03.598 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T20:40:03.598 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T20:40:03.598 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T20:40:03.624 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T20:40:03.625 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:03.625 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T20:40:03.625 INFO:teuthology.orchestra.run.vm03.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T20:40:03.625 INFO:teuthology.orchestra.run.vm03.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T20:40:03.625 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:03.625 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:03.631 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T20:40:03.631 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:03.631 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T20:40:03.631 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T20:40:03.631 INFO:teuthology.orchestra.run.vm04.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T20:40:03.631 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:03.631 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:03.654 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T20:40:03.655 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:03.655 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:40:03.655 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T20:40:03.655 INFO:teuthology.orchestra.run.vm08.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T20:40:03.655 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:03.655 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:03.857 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:03.857 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout:Removing unused dependencies: 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout:Remove 8 Packages 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 28 M 2026-03-09T20:40:03.858 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T20:40:03.861 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T20:40:03.861 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T20:40:03.863 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout:Remove 8 Packages 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 28 M 2026-03-09T20:40:03.864 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T20:40:03.867 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T20:40:03.867 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T20:40:03.875 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout:Remove 8 Packages 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 28 M 2026-03-09T20:40:03.876 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:40:03.879 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:40:03.879 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:40:03.884 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T20:40:03.884 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T20:40:03.890 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T20:40:03.890 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T20:40:03.902 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:40:03.902 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:40:03.927 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T20:40:03.931 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T20:40:03.932 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T20:40:03.935 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T20:40:03.935 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T20:40:03.937 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T20:40:03.939 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T20:40:03.939 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T20:40:03.941 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T20:40:03.942 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T20:40:03.943 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T20:40:03.944 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T20:40:03.946 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T20:40:03.947 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:40:03.948 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T20:40:03.952 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T20:40:03.956 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T20:40:03.958 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T20:40:03.961 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T20:40:03.964 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T20:40:03.965 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:40:03.965 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:03.965 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T20:40:03.966 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T20:40:03.966 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T20:40:03.966 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:03.966 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T20:40:03.966 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:40:03.971 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:40:03.971 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:03.971 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T20:40:03.971 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T20:40:03.971 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T20:40:03.971 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:03.972 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:40:03.973 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:40:03.979 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:40:03.987 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:40:03.987 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:03.987 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T20:40:03.987 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T20:40:03.987 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T20:40:03.987 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:03.988 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:40:03.994 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:40:03.994 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:03.994 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T20:40:03.994 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T20:40:03.994 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T20:40:03.994 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:03.995 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T20:40:03.995 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:40:03.999 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:40:03.999 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:03.999 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T20:40:03.999 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T20:40:03.999 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T20:40:03.999 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:04.001 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:40:04.019 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:40:04.019 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:04.019 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T20:40:04.019 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T20:40:04.019 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T20:40:04.019 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:04.019 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:40:04.084 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:40:04.084 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T20:40:04.084 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T20:40:04.084 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T20:40:04.084 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T20:40:04.084 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T20:40:04.084 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T20:40:04.084 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T20:40:04.089 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:40:04.089 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T20:40:04.089 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T20:40:04.089 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T20:40:04.089 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T20:40:04.089 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T20:40:04.089 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T20:40:04.089 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T20:40:04.116 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T20:40:04.116 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T20:40:04.116 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T20:40:04.116 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T20:40:04.116 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T20:40:04.116 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T20:40:04.116 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T20:40:04.116 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T20:40:04.140 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T20:40:04.140 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:04.140 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T20:40:04.140 INFO:teuthology.orchestra.run.vm04.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:04.140 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:04.140 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:04.140 INFO:teuthology.orchestra.run.vm04.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T20:40:04.140 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T20:40:04.140 INFO:teuthology.orchestra.run.vm04.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T20:40:04.140 INFO:teuthology.orchestra.run.vm04.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T20:40:04.140 INFO:teuthology.orchestra.run.vm04.stdout: zip-3.0-35.el9.x86_64 2026-03-09T20:40:04.140 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:04.140 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:04.153 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T20:40:04.153 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:04.153 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T20:40:04.153 INFO:teuthology.orchestra.run.vm03.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:04.153 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:04.153 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:04.153 INFO:teuthology.orchestra.run.vm03.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T20:40:04.153 INFO:teuthology.orchestra.run.vm03.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T20:40:04.153 INFO:teuthology.orchestra.run.vm03.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T20:40:04.153 INFO:teuthology.orchestra.run.vm03.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T20:40:04.153 INFO:teuthology.orchestra.run.vm03.stdout: zip-3.0-35.el9.x86_64 2026-03-09T20:40:04.153 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:04.153 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:04.168 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T20:40:04.168 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:04.168 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:40:04.168 INFO:teuthology.orchestra.run.vm08.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:04.168 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:04.168 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:04.168 INFO:teuthology.orchestra.run.vm08.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T20:40:04.168 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T20:40:04.168 INFO:teuthology.orchestra.run.vm08.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T20:40:04.168 INFO:teuthology.orchestra.run.vm08.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T20:40:04.168 INFO:teuthology.orchestra.run.vm08.stdout: zip-3.0-35.el9.x86_64 2026-03-09T20:40:04.168 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:04.168 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:04.373 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout:=========================================================================================== 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout:=========================================================================================== 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T20:40:04.379 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T20:40:04.380 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout:=========================================================================================== 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout:Remove 102 Packages 2026-03-09T20:40:04.381 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:04.382 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 613 M 2026-03-09T20:40:04.382 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T20:40:04.396 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:04.398 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:04.402 INFO:teuthology.orchestra.run.vm03.stdout:=========================================================================================== 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout:=========================================================================================== 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout:Removing dependent packages: 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout:Removing unused dependencies: 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T20:40:04.403 INFO:teuthology.orchestra.run.vm03.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T20:40:04.404 INFO:teuthology.orchestra.run.vm03.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout:=========================================================================================== 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout:Remove 102 Packages 2026-03-09T20:40:04.405 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 613 M 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout:=========================================================================================== 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout:=========================================================================================== 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout:Removing dependent packages: 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T20:40:04.406 INFO:teuthology.orchestra.run.vm08.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T20:40:04.407 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout:=========================================================================================== 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout:Remove 102 Packages 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 613 M 2026-03-09T20:40:04.408 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:40:04.409 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T20:40:04.409 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T20:40:04.432 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T20:40:04.432 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T20:40:04.432 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:40:04.432 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:40:04.511 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T20:40:04.512 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T20:40:04.534 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:40:04.534 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:40:04.538 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T20:40:04.538 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T20:40:04.657 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T20:40:04.657 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T20:40:04.664 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T20:40:04.678 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:40:04.678 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T20:40:04.682 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:40:04.682 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:04.682 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T20:40:04.682 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T20:40:04.682 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T20:40:04.683 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:04.683 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:40:04.685 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T20:40:04.686 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T20:40:04.686 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T20:40:04.694 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T20:40:04.696 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:40:04.702 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:40:04.702 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:04.702 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T20:40:04.702 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T20:40:04.702 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T20:40:04.703 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:04.703 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:40:04.714 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:40:04.714 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:04.714 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T20:40:04.714 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T20:40:04.714 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T20:40:04.714 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:04.714 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:40:04.716 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:40:04.720 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T20:40:04.720 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T20:40:04.730 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:40:04.740 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T20:40:04.740 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T20:40:04.756 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T20:40:04.756 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T20:40:04.776 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T20:40:04.785 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T20:40:04.789 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T20:40:04.789 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:40:04.796 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T20:40:04.801 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:40:04.805 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T20:40:04.808 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T20:40:04.809 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T20:40:04.809 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:40:04.809 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T20:40:04.812 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T20:40:04.818 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T20:40:04.821 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:40:04.822 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T20:40:04.822 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T20:40:04.822 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:40:04.826 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T20:40:04.827 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T20:40:04.831 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T20:40:04.834 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:40:04.842 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T20:40:04.844 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T20:40:04.846 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T20:40:04.848 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:40:04.848 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:04.848 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T20:40:04.848 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T20:40:04.848 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T20:40:04.848 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:04.849 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T20:40:04.855 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T20:40:04.855 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:40:04.859 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T20:40:04.865 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:40:04.868 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:40:04.868 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:04.868 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T20:40:04.868 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T20:40:04.868 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T20:40:04.868 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:04.874 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:40:04.880 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:40:04.880 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:04.880 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T20:40:04.881 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T20:40:04.881 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T20:40:04.881 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:04.882 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:40:04.883 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:04.883 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T20:40:04.883 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:04.883 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:40:04.886 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:40:04.892 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:40:04.895 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:40:04.899 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:40:04.899 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:04.899 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T20:40:04.900 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:04.902 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:40:04.905 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T20:40:04.908 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:40:04.910 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T20:40:04.914 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:40:04.914 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:04.914 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T20:40:04.914 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:04.915 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T20:40:04.918 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:40:04.920 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T20:40:04.922 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:40:04.924 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T20:40:04.925 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T20:40:04.930 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T20:40:04.932 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T20:40:04.934 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T20:40:04.937 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T20:40:04.938 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T20:40:04.939 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T20:40:04.944 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T20:40:04.944 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T20:40:04.952 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T20:40:04.953 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T20:40:04.955 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T20:40:04.960 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T20:40:04.962 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T20:40:04.964 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T20:40:04.970 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T20:40:04.970 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T20:40:04.976 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T20:40:04.980 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T20:40:04.986 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T20:40:04.993 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T20:40:05.000 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T20:40:05.003 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T20:40:05.008 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T20:40:05.012 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T20:40:05.015 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T20:40:05.016 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T20:40:05.019 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T20:40:05.022 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T20:40:05.023 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T20:40:05.024 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T20:40:05.025 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T20:40:05.028 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T20:40:05.032 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T20:40:05.036 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T20:40:05.039 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T20:40:05.039 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T20:40:05.045 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T20:40:05.047 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T20:40:05.047 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T20:40:05.054 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T20:40:05.124 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T20:40:05.137 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T20:40:05.140 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T20:40:05.146 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T20:40:05.153 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T20:40:05.155 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:40:05.155 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T20:40:05.155 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:05.157 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:40:05.161 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T20:40:05.167 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:40:05.167 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T20:40:05.167 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:05.168 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:40:05.174 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:40:05.174 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T20:40:05.174 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:05.175 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:40:05.185 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:40:05.196 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:40:05.201 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T20:40:05.201 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T20:40:05.207 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T20:40:05.210 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T20:40:05.211 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T20:40:05.212 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T20:40:05.216 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T20:40:05.217 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T20:40:05.219 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T20:40:05.221 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T20:40:05.223 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T20:40:05.225 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T20:40:05.228 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T20:40:05.233 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:40:05.234 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:05.234 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T20:40:05.234 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T20:40:05.234 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T20:40:05.234 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:05.235 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:40:05.239 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:40:05.239 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:05.239 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T20:40:05.239 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T20:40:05.239 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T20:40:05.239 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:05.241 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:40:05.249 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:40:05.249 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:40:05.249 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:05.249 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T20:40:05.249 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T20:40:05.249 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T20:40:05.249 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:05.251 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:40:05.252 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:40:05.253 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T20:40:05.255 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T20:40:05.255 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T20:40:05.258 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T20:40:05.258 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T20:40:05.260 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T20:40:05.262 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T20:40:05.263 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T20:40:05.264 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T20:40:05.265 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T20:40:05.266 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T20:40:05.268 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T20:40:05.269 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T20:40:05.270 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T20:40:05.270 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T20:40:05.273 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T20:40:05.274 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T20:40:05.274 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T20:40:05.276 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T20:40:05.279 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T20:40:05.283 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T20:40:05.288 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T20:40:05.321 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T20:40:05.327 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T20:40:05.333 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T20:40:05.336 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T20:40:05.336 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T20:40:05.339 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T20:40:05.342 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T20:40:05.342 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T20:40:05.344 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T20:40:05.347 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T20:40:05.348 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T20:40:05.348 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T20:40:05.350 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T20:40:05.350 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T20:40:05.351 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T20:40:05.353 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T20:40:05.355 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T20:40:05.356 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T20:40:05.357 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T20:40:05.360 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T20:40:05.363 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T20:40:05.371 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:40:05.371 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:05.371 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T20:40:05.371 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:05.372 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:40:05.379 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:40:05.381 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T20:40:05.381 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:40:05.381 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:05.381 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T20:40:05.381 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:05.382 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:40:05.383 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T20:40:05.386 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T20:40:05.386 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:40:05.386 INFO:teuthology.orchestra.run.vm03.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T20:40:05.386 INFO:teuthology.orchestra.run.vm03.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T20:40:05.386 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:05.387 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:40:05.388 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T20:40:05.390 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:40:05.390 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T20:40:05.392 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T20:40:05.393 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T20:40:05.394 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T20:40:05.396 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T20:40:05.396 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T20:40:05.397 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T20:40:05.398 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T20:40:05.399 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T20:40:05.399 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T20:40:05.400 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T20:40:05.401 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T20:40:05.403 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T20:40:05.404 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T20:40:05.406 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T20:40:05.406 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T20:40:05.407 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T20:40:05.408 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T20:40:05.410 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T20:40:05.411 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T20:40:05.412 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T20:40:05.414 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T20:40:05.414 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T20:40:05.416 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T20:40:05.417 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T20:40:05.417 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T20:40:05.419 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T20:40:05.422 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T20:40:05.424 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T20:40:05.424 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T20:40:05.425 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T20:40:05.426 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T20:40:05.429 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T20:40:05.429 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T20:40:05.430 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T20:40:05.431 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T20:40:05.434 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T20:40:05.435 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T20:40:05.436 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T20:40:05.437 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T20:40:05.439 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T20:40:05.441 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T20:40:05.442 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T20:40:05.445 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T20:40:05.446 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T20:40:05.447 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T20:40:05.450 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T20:40:05.451 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T20:40:05.452 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T20:40:05.454 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T20:40:05.456 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T20:40:05.456 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T20:40:05.457 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T20:40:05.460 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T20:40:05.462 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T20:40:05.462 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T20:40:05.464 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T20:40:05.465 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T20:40:05.465 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T20:40:05.466 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T20:40:05.469 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T20:40:05.469 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T20:40:05.472 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T20:40:05.472 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T20:40:05.476 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T20:40:05.477 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T20:40:05.477 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T20:40:05.479 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T20:40:05.481 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T20:40:05.483 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T20:40:05.485 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T20:40:05.486 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T20:40:05.487 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T20:40:05.489 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T20:40:05.490 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T20:40:05.492 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T20:40:05.493 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T20:40:05.496 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T20:40:05.496 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T20:40:05.498 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T20:40:05.499 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T20:40:05.500 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T20:40:05.500 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T20:40:05.502 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T20:40:05.504 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T20:40:05.506 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T20:40:05.506 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T20:40:05.509 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T20:40:05.511 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T20:40:05.515 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T20:40:05.519 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:40:05.519 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T20:40:05.519 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:05.526 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:40:05.531 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:40:05.532 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T20:40:05.532 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:05.538 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:40:05.538 INFO:teuthology.orchestra.run.vm03.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T20:40:05.538 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:05.538 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:40:05.546 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:40:05.553 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:40:05.553 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T20:40:05.565 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T20:40:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:40:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T20:40:05.570 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T20:40:05.573 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T20:40:05.575 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T20:40:05.575 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T20:40:05.575 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T20:40:05.575 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T20:40:05.581 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T20:40:05.586 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T20:40:05.587 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T20:40:05.588 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T20:40:05.590 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T20:40:05.590 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T20:40:05.592 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T20:40:05.595 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T20:40:05.597 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T20:40:05.597 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T20:40:11.329 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T20:40:11.329 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /sys 2026-03-09T20:40:11.329 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /proc 2026-03-09T20:40:11.329 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /mnt 2026-03-09T20:40:11.329 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /var/tmp 2026-03-09T20:40:11.329 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /home 2026-03-09T20:40:11.329 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /root 2026-03-09T20:40:11.329 INFO:teuthology.orchestra.run.vm03.stdout:skipping the directory /tmp 2026-03-09T20:40:11.329 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:11.334 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T20:40:11.335 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /sys 2026-03-09T20:40:11.335 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /proc 2026-03-09T20:40:11.335 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /mnt 2026-03-09T20:40:11.335 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /var/tmp 2026-03-09T20:40:11.335 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /home 2026-03-09T20:40:11.335 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /root 2026-03-09T20:40:11.335 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /tmp 2026-03-09T20:40:11.335 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:11.339 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T20:40:11.344 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T20:40:11.360 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:40:11.361 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:40:11.365 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:40:11.365 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:40:11.368 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:40:11.371 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T20:40:11.374 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:40:11.374 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T20:40:11.376 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T20:40:11.377 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T20:40:11.378 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T20:40:11.378 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /sys 2026-03-09T20:40:11.378 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /proc 2026-03-09T20:40:11.378 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /mnt 2026-03-09T20:40:11.378 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /var/tmp 2026-03-09T20:40:11.378 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /home 2026-03-09T20:40:11.378 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /root 2026-03-09T20:40:11.378 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /tmp 2026-03-09T20:40:11.378 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:11.378 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T20:40:11.378 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T20:40:11.379 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T20:40:11.383 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T20:40:11.385 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T20:40:11.385 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T20:40:11.387 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T20:40:11.393 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T20:40:11.394 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T20:40:11.397 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T20:40:11.398 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T20:40:11.400 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T20:40:11.400 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T20:40:11.403 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T20:40:11.403 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T20:40:11.405 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:40:11.405 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:40:11.406 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T20:40:11.408 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T20:40:11.408 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T20:40:11.412 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T20:40:11.413 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T20:40:11.415 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T20:40:11.415 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T20:40:11.417 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T20:40:11.420 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T20:40:11.420 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T20:40:11.420 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:40:11.421 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T20:40:11.422 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T20:40:11.422 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T20:40:11.424 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T20:40:11.425 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:40:11.436 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T20:40:11.437 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T20:40:11.439 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T20:40:11.442 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T20:40:11.444 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T20:40:11.449 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T20:40:11.457 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T20:40:11.461 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T20:40:11.461 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:40:11.524 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:40:11.524 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T20:40:11.524 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:40:11.524 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T20:40:11.524 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T20:40:11.524 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T20:40:11.524 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T20:40:11.524 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:40:11.524 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T20:40:11.524 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T20:40:11.526 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T20:40:11.527 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T20:40:11.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T20:40:11.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T20:40:11.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T20:40:11.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T20:40:11.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T20:40:11.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T20:40:11.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T20:40:11.528 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T20:40:11.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T20:40:11.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T20:40:11.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T20:40:11.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T20:40:11.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T20:40:11.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T20:40:11.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T20:40:11.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T20:40:11.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T20:40:11.529 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T20:40:11.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T20:40:11.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T20:40:11.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T20:40:11.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T20:40:11.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T20:40:11.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T20:40:11.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T20:40:11.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T20:40:11.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T20:40:11.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T20:40:11.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T20:40:11.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T20:40:11.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T20:40:11.530 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T20:40:11.532 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T20:40:11.533 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T20:40:11.534 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T20:40:11.564 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T20:40:11.565 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T20:40:11.566 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T20:40:11.607 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:40:11.607 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:11.607 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T20:40:11.607 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T20:40:11.607 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.607 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.607 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T20:40:11.608 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T20:40:11.609 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm03.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm03.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm03.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm03.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm03.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.610 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T20:40:11.611 INFO:teuthology.orchestra.run.vm03.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T20:40:11.612 INFO:teuthology.orchestra.run.vm03.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T20:40:11.613 INFO:teuthology.orchestra.run.vm03.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T20:40:11.613 INFO:teuthology.orchestra.run.vm03.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T20:40:11.613 INFO:teuthology.orchestra.run.vm03.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T20:40:11.613 INFO:teuthology.orchestra.run.vm03.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T20:40:11.613 INFO:teuthology.orchestra.run.vm03.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T20:40:11.613 INFO:teuthology.orchestra.run.vm03.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T20:40:11.613 INFO:teuthology.orchestra.run.vm03.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T20:40:11.613 INFO:teuthology.orchestra.run.vm03.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T20:40:11.613 INFO:teuthology.orchestra.run.vm03.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.613 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:11.613 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:11.651 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T20:40:11.652 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T20:40:11.653 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:11.654 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:11.831 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:11.831 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:11.831 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T20:40:11.831 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:11.831 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T20:40:11.831 INFO:teuthology.orchestra.run.vm03.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T20:40:11.831 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:11.831 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T20:40:11.831 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:11.831 INFO:teuthology.orchestra.run.vm03.stdout:Remove 1 Package 2026-03-09T20:40:11.831 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:11.831 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 775 k 2026-03-09T20:40:11.832 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T20:40:11.833 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T20:40:11.833 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T20:40:11.833 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:11.834 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:11.834 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T20:40:11.834 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:11.834 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T20:40:11.834 INFO:teuthology.orchestra.run.vm04.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T20:40:11.834 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:11.834 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T20:40:11.834 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:11.834 INFO:teuthology.orchestra.run.vm04.stdout:Remove 1 Package 2026-03-09T20:40:11.834 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:11.834 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 775 k 2026-03-09T20:40:11.834 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T20:40:11.834 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T20:40:11.834 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T20:40:11.836 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T20:40:11.836 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T20:40:11.837 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T20:40:11.837 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T20:40:11.851 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T20:40:11.851 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:40:11.854 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T20:40:11.854 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:40:11.862 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:11.863 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:11.863 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:40:11.863 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:11.863 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:40:11.863 INFO:teuthology.orchestra.run.vm08.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T20:40:11.863 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:11.863 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:40:11.863 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:11.863 INFO:teuthology.orchestra.run.vm08.stdout:Remove 1 Package 2026-03-09T20:40:11.863 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:11.863 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 775 k 2026-03-09T20:40:11.863 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:40:11.865 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:40:11.865 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:40:11.866 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:40:11.866 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:40:11.888 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:40:11.888 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:40:11.966 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:40:11.968 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:40:12.017 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:40:12.017 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:12.017 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T20:40:12.017 INFO:teuthology.orchestra.run.vm03.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:12.017 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:12.017 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:12.033 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:40:12.038 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:40:12.038 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:12.038 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T20:40:12.038 INFO:teuthology.orchestra.run.vm04.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:12.038 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:12.038 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:12.083 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T20:40:12.083 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:12.083 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:40:12.083 INFO:teuthology.orchestra.run.vm08.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T20:40:12.083 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:12.083 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:12.207 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T20:40:12.208 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:12.211 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:12.212 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:12.212 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:12.224 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T20:40:12.224 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:12.228 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:12.229 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:12.229 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:12.286 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T20:40:12.286 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:12.289 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:12.290 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:12.290 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:12.405 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: ceph-mgr 2026-03-09T20:40:12.405 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:12.407 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr 2026-03-09T20:40:12.407 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:12.409 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:12.409 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:12.409 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:12.410 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:12.411 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:12.411 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:12.474 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr 2026-03-09T20:40:12.474 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:12.478 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:12.478 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:12.478 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:12.593 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T20:40:12.594 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:12.595 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T20:40:12.596 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:12.596 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:12.597 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:12.597 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:12.599 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:12.600 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:12.600 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:12.671 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T20:40:12.671 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:12.674 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:12.675 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:12.675 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:12.770 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T20:40:12.770 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:12.772 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T20:40:12.772 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:12.774 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:12.774 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:12.774 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:12.776 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:12.776 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:12.776 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:12.855 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T20:40:12.855 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:12.858 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:12.859 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:12.859 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:12.950 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-rook 2026-03-09T20:40:12.951 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:12.954 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:12.954 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:12.954 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:12.960 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: ceph-mgr-rook 2026-03-09T20:40:12.960 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:12.963 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:12.963 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:12.963 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:13.039 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr-rook 2026-03-09T20:40:13.039 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:13.043 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:13.043 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:13.043 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:13.127 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T20:40:13.127 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:13.130 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:13.131 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:13.131 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:13.141 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T20:40:13.141 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:13.145 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:13.145 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:13.145 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:13.234 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T20:40:13.234 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:13.238 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:13.238 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:13.238 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:13.317 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:13.317 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:13.317 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T20:40:13.317 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:13.317 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T20:40:13.317 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T20:40:13.317 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:13.317 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T20:40:13.317 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:13.317 INFO:teuthology.orchestra.run.vm04.stdout:Remove 1 Package 2026-03-09T20:40:13.317 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:13.317 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 3.6 M 2026-03-09T20:40:13.317 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T20:40:13.319 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T20:40:13.319 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T20:40:13.329 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T20:40:13.329 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T20:40:13.337 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:13.337 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:13.338 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T20:40:13.338 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:13.338 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T20:40:13.338 INFO:teuthology.orchestra.run.vm03.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T20:40:13.338 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:13.338 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T20:40:13.338 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:13.338 INFO:teuthology.orchestra.run.vm03.stdout:Remove 1 Package 2026-03-09T20:40:13.338 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:13.338 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 3.6 M 2026-03-09T20:40:13.338 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T20:40:13.339 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T20:40:13.340 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T20:40:13.350 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T20:40:13.350 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T20:40:13.354 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T20:40:13.368 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:40:13.375 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T20:40:13.389 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:40:13.431 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:13.432 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:13.432 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:40:13.432 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:13.432 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:40:13.432 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T20:40:13.432 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:13.432 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:40:13.432 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:13.432 INFO:teuthology.orchestra.run.vm08.stdout:Remove 1 Package 2026-03-09T20:40:13.432 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:13.432 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 3.6 M 2026-03-09T20:40:13.432 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:40:13.434 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:40:13.434 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:40:13.434 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:40:13.444 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:40:13.444 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:40:13.453 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:40:13.480 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:40:13.487 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:40:13.487 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:13.487 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T20:40:13.487 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:13.487 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:13.487 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:13.498 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:40:13.507 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:40:13.507 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:13.507 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T20:40:13.507 INFO:teuthology.orchestra.run.vm03.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:13.507 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:13.507 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:13.559 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:40:13.609 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T20:40:13.609 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:13.609 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:40:13.609 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:13.609 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:13.609 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:13.679 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-volume 2026-03-09T20:40:13.679 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:13.683 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:13.683 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:13.683 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:13.689 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: ceph-volume 2026-03-09T20:40:13.689 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:13.692 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:13.693 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:13.693 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:13.807 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: ceph-volume 2026-03-09T20:40:13.808 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:13.811 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:13.811 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:13.811 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:13.891 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:13.891 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:13.891 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repo Size 2026-03-09T20:40:13.891 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:13.891 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T20:40:13.891 INFO:teuthology.orchestra.run.vm03.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T20:40:13.891 INFO:teuthology.orchestra.run.vm03.stdout:Removing dependent packages: 2026-03-09T20:40:13.891 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T20:40:13.891 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:13.891 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T20:40:13.891 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:13.891 INFO:teuthology.orchestra.run.vm03.stdout:Remove 2 Packages 2026-03-09T20:40:13.891 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:13.891 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 610 k 2026-03-09T20:40:13.892 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T20:40:13.893 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T20:40:13.893 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T20:40:13.902 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:13.902 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:13.902 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repo Size 2026-03-09T20:40:13.902 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:13.902 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T20:40:13.902 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T20:40:13.902 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-09T20:40:13.902 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T20:40:13.902 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:13.902 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T20:40:13.902 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:13.902 INFO:teuthology.orchestra.run.vm04.stdout:Remove 2 Packages 2026-03-09T20:40:13.902 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:13.903 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 610 k 2026-03-09T20:40:13.903 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T20:40:13.904 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T20:40:13.904 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T20:40:13.904 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T20:40:13.905 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T20:40:13.915 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T20:40:13.916 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T20:40:13.931 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T20:40:13.933 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:13.941 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T20:40:13.944 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:13.946 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:40:13.957 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:40:14.015 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:40:14.015 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:14.017 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:14.019 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:14.019 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repo Size 2026-03-09T20:40:14.019 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:14.019 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:40:14.019 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T20:40:14.019 INFO:teuthology.orchestra.run.vm08.stdout:Removing dependent packages: 2026-03-09T20:40:14.019 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T20:40:14.019 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:14.019 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:40:14.019 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:14.019 INFO:teuthology.orchestra.run.vm08.stdout:Remove 2 Packages 2026-03-09T20:40:14.019 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:14.019 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 610 k 2026-03-09T20:40:14.019 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:40:14.021 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:40:14.021 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:40:14.032 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:40:14.032 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:40:14.033 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:40:14.033 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:14.059 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:40:14.061 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:14.067 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:40:14.067 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:14.067 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T20:40:14.067 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.067 INFO:teuthology.orchestra.run.vm03.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.067 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:14.067 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:14.075 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:40:14.082 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:40:14.082 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:14.082 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T20:40:14.083 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.083 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.083 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:14.083 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:14.135 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:40:14.135 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T20:40:14.183 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T20:40:14.183 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:14.183 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:40:14.183 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.183 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.183 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:14.183 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:14.270 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:14.270 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:14.270 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repo Size 2026-03-09T20:40:14.270 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:14.271 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T20:40:14.271 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T20:40:14.271 INFO:teuthology.orchestra.run.vm03.stdout:Removing dependent packages: 2026-03-09T20:40:14.271 INFO:teuthology.orchestra.run.vm03.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T20:40:14.271 INFO:teuthology.orchestra.run.vm03.stdout:Removing unused dependencies: 2026-03-09T20:40:14.271 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T20:40:14.271 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:14.271 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T20:40:14.271 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:14.271 INFO:teuthology.orchestra.run.vm03.stdout:Remove 3 Packages 2026-03-09T20:40:14.271 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:14.271 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 3.7 M 2026-03-09T20:40:14.271 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T20:40:14.273 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T20:40:14.273 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T20:40:14.288 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:14.288 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:14.288 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repo Size 2026-03-09T20:40:14.288 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:14.288 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T20:40:14.288 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T20:40:14.288 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-09T20:40:14.289 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T20:40:14.289 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T20:40:14.289 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T20:40:14.289 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:14.289 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T20:40:14.289 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:14.289 INFO:teuthology.orchestra.run.vm04.stdout:Remove 3 Packages 2026-03-09T20:40:14.289 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:14.289 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 3.7 M 2026-03-09T20:40:14.289 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T20:40:14.289 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T20:40:14.289 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T20:40:14.291 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T20:40:14.291 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T20:40:14.306 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T20:40:14.306 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T20:40:14.352 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T20:40:14.352 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T20:40:14.355 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T20:40:14.356 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T20:40:14.365 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T20:40:14.365 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:40:14.369 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T20:40:14.369 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:40:14.374 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repo Size 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout:Removing dependent packages: 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout:Remove 3 Packages 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 3.7 M 2026-03-09T20:40:14.375 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:40:14.377 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:40:14.377 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:40:14.394 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:40:14.394 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:40:14.437 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:40:14.439 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T20:40:14.440 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T20:40:14.440 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:40:14.459 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:40:14.459 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T20:40:14.459 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T20:40:14.460 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:40:14.460 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T20:40:14.460 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T20:40:14.496 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:40:14.496 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:14.496 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T20:40:14.496 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.496 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.496 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.496 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:14.496 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:14.497 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:40:14.498 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:14.498 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T20:40:14.498 INFO:teuthology.orchestra.run.vm03.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.498 INFO:teuthology.orchestra.run.vm03.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.498 INFO:teuthology.orchestra.run.vm03.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.498 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:14.498 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:14.504 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:40:14.504 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T20:40:14.504 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T20:40:14.542 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T20:40:14.542 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:14.542 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:40:14.542 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.542 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.542 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:14.542 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:14.542 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:14.673 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: libcephfs-devel 2026-03-09T20:40:14.673 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:14.676 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:14.676 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:14.677 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:14.679 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: libcephfs-devel 2026-03-09T20:40:14.679 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:14.682 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:14.683 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:14.683 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:14.721 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: libcephfs-devel 2026-03-09T20:40:14.721 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:14.724 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:14.725 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:14.725 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:14.866 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:14.867 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:14.867 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:14.867 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T20:40:14.867 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:14.867 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout:Remove 20 Packages 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 79 M 2026-03-09T20:40:14.868 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: Package Arch Version Repository Size 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout:Removing: 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout:Removing dependent packages: 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout:Removing unused dependencies: 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout:Transaction Summary 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout:================================================================================ 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout:Remove 20 Packages 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout:Freed space: 79 M 2026-03-09T20:40:14.869 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction check 2026-03-09T20:40:14.872 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T20:40:14.872 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T20:40:14.873 INFO:teuthology.orchestra.run.vm03.stdout:Transaction check succeeded. 2026-03-09T20:40:14.873 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction test 2026-03-09T20:40:14.894 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T20:40:14.894 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T20:40:14.895 INFO:teuthology.orchestra.run.vm03.stdout:Transaction test succeeded. 2026-03-09T20:40:14.895 INFO:teuthology.orchestra.run.vm03.stdout:Running transaction 2026-03-09T20:40:14.910 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout:Removing dependent packages: 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout:Remove 20 Packages 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 79 M 2026-03-09T20:40:14.912 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-03-09T20:40:14.916 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-03-09T20:40:14.916 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-03-09T20:40:14.936 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T20:40:14.937 INFO:teuthology.orchestra.run.vm03.stdout: Preparing : 1/1 2026-03-09T20:40:14.938 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T20:40:14.939 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-03-09T20:40:14.939 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-03-09T20:40:14.940 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T20:40:14.941 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T20:40:14.942 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T20:40:14.944 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T20:40:14.944 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T20:40:14.945 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T20:40:14.945 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T20:40:14.956 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T20:40:14.958 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T20:40:14.958 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T20:40:14.960 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T20:40:14.960 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T20:40:14.962 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T20:40:14.962 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T20:40:14.963 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T20:40:14.964 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T20:40:14.965 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T20:40:14.966 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T20:40:14.966 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:40:14.967 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T20:40:14.967 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:40:14.981 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-03-09T20:40:14.983 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:40:14.983 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T20:40:14.983 INFO:teuthology.orchestra.run.vm03.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T20:40:14.983 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:14.984 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T20:40:14.986 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T20:40:14.987 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:40:14.987 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T20:40:14.987 INFO:teuthology.orchestra.run.vm04.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T20:40:14.987 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:14.989 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T20:40:14.989 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T20:40:14.998 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T20:40:15.000 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T20:40:15.002 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T20:40:15.003 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T20:40:15.004 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T20:40:15.005 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T20:40:15.008 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T20:40:15.008 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T20:40:15.008 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T20:40:15.010 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T20:40:15.011 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T20:40:15.013 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T20:40:15.013 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T20:40:15.013 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T20:40:15.016 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T20:40:15.016 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T20:40:15.016 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T20:40:15.018 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T20:40:15.018 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T20:40:15.020 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T20:40:15.020 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T20:40:15.020 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:40:15.021 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T20:40:15.023 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T20:40:15.026 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T20:40:15.033 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:40:15.033 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T20:40:15.033 INFO:teuthology.orchestra.run.vm08.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T20:40:15.033 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:15.034 INFO:teuthology.orchestra.run.vm03.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T20:40:15.041 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T20:40:15.053 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T20:40:15.055 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T20:40:15.059 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T20:40:15.062 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T20:40:15.065 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T20:40:15.068 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T20:40:15.069 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T20:40:15.071 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T20:40:15.077 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T20:40:15.093 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T20:40:15.098 INFO:teuthology.orchestra.run.vm03.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T20:40:15.098 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T20:40:15.098 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T20:40:15.098 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T20:40:15.098 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T20:40:15.098 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T20:40:15.098 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T20:40:15.098 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T20:40:15.099 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T20:40:15.099 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T20:40:15.099 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:40:15.099 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T20:40:15.099 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T20:40:15.099 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T20:40:15.099 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T20:40:15.099 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T20:40:15.099 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T20:40:15.099 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T20:40:15.099 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T20:40:15.099 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T20:40:15.101 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T20:40:15.101 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T20:40:15.101 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T20:40:15.102 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout:Removed: 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout: 2026-03-09T20:40:15.140 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:15.149 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T20:40:15.149 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:15.149 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T20:40:15.149 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T20:40:15.149 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T20:40:15.149 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T20:40:15.149 INFO:teuthology.orchestra.run.vm04.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T20:40:15.149 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T20:40:15.149 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T20:40:15.149 INFO:teuthology.orchestra.run.vm04.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T20:40:15.150 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T20:40:15.167 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T20:40:15.207 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T20:40:15.207 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T20:40:15.208 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:15.357 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: librbd1 2026-03-09T20:40:15.357 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:15.359 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:15.360 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:15.360 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:15.392 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: librbd1 2026-03-09T20:40:15.392 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:15.394 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:15.395 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:15.395 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:15.432 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: librbd1 2026-03-09T20:40:15.433 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:15.435 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:15.435 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:15.435 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:15.566 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: python3-rados 2026-03-09T20:40:15.566 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:15.569 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:15.569 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:15.569 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:15.596 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rados 2026-03-09T20:40:15.596 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:15.598 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:15.599 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:15.599 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:15.648 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: python3-rados 2026-03-09T20:40:15.649 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:15.651 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:15.651 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:15.652 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:15.756 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: python3-rgw 2026-03-09T20:40:15.756 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:15.758 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:15.759 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:15.759 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:15.763 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rgw 2026-03-09T20:40:15.763 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:15.765 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:15.766 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:15.766 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:15.833 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: python3-rgw 2026-03-09T20:40:15.833 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:15.836 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:15.836 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:15.836 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:15.930 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-cephfs 2026-03-09T20:40:15.930 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:15.933 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:15.933 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:15.933 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:15.937 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: python3-cephfs 2026-03-09T20:40:15.937 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:15.939 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:15.939 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:15.939 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:16.006 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: python3-cephfs 2026-03-09T20:40:16.006 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:16.008 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:16.009 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:16.009 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:16.123 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rbd 2026-03-09T20:40:16.124 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:16.125 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:16.126 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:16.126 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:16.126 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: python3-rbd 2026-03-09T20:40:16.126 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:16.129 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:16.129 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:16.129 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:16.244 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: python3-rbd 2026-03-09T20:40:16.244 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:16.246 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:16.247 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:16.247 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:16.318 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-fuse 2026-03-09T20:40:16.319 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:16.320 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:16.321 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:16.321 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:16.328 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: rbd-fuse 2026-03-09T20:40:16.329 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:16.330 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:16.331 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:16.331 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:16.452 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: rbd-fuse 2026-03-09T20:40:16.452 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:16.454 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:16.455 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:16.455 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:16.566 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: rbd-mirror 2026-03-09T20:40:16.566 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:16.568 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:16.569 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:16.569 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:16.581 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-mirror 2026-03-09T20:40:16.581 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:16.583 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:16.584 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:16.584 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:16.664 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: rbd-mirror 2026-03-09T20:40:16.664 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:16.667 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:16.667 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:16.667 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:16.744 INFO:teuthology.orchestra.run.vm03.stdout:No match for argument: rbd-nbd 2026-03-09T20:40:16.744 INFO:teuthology.orchestra.run.vm03.stderr:No packages marked for removal. 2026-03-09T20:40:16.746 INFO:teuthology.orchestra.run.vm03.stdout:Dependencies resolved. 2026-03-09T20:40:16.746 INFO:teuthology.orchestra.run.vm03.stdout:Nothing to do. 2026-03-09T20:40:16.746 INFO:teuthology.orchestra.run.vm03.stdout:Complete! 2026-03-09T20:40:16.759 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-nbd 2026-03-09T20:40:16.759 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T20:40:16.761 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T20:40:16.762 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T20:40:16.762 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T20:40:16.765 DEBUG:teuthology.orchestra.run.vm03:> sudo yum clean all 2026-03-09T20:40:16.790 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean all 2026-03-09T20:40:16.844 INFO:teuthology.orchestra.run.vm08.stdout:No match for argument: rbd-nbd 2026-03-09T20:40:16.844 INFO:teuthology.orchestra.run.vm08.stderr:No packages marked for removal. 2026-03-09T20:40:16.846 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-03-09T20:40:16.847 INFO:teuthology.orchestra.run.vm08.stdout:Nothing to do. 2026-03-09T20:40:16.847 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-03-09T20:40:16.870 DEBUG:teuthology.orchestra.run.vm08:> sudo yum clean all 2026-03-09T20:40:16.902 INFO:teuthology.orchestra.run.vm03.stdout:56 files removed 2026-03-09T20:40:16.919 INFO:teuthology.orchestra.run.vm04.stdout:56 files removed 2026-03-09T20:40:16.928 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T20:40:16.947 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T20:40:16.957 DEBUG:teuthology.orchestra.run.vm03:> sudo yum clean expire-cache 2026-03-09T20:40:16.973 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean expire-cache 2026-03-09T20:40:17.003 INFO:teuthology.orchestra.run.vm08.stdout:56 files removed 2026-03-09T20:40:17.030 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T20:40:17.058 DEBUG:teuthology.orchestra.run.vm08:> sudo yum clean expire-cache 2026-03-09T20:40:17.112 INFO:teuthology.orchestra.run.vm03.stdout:Cache was expired 2026-03-09T20:40:17.112 INFO:teuthology.orchestra.run.vm03.stdout:0 files removed 2026-03-09T20:40:17.128 INFO:teuthology.orchestra.run.vm04.stdout:Cache was expired 2026-03-09T20:40:17.128 INFO:teuthology.orchestra.run.vm04.stdout:0 files removed 2026-03-09T20:40:17.129 DEBUG:teuthology.parallel:result is None 2026-03-09T20:40:17.145 DEBUG:teuthology.parallel:result is None 2026-03-09T20:40:17.220 INFO:teuthology.orchestra.run.vm08.stdout:Cache was expired 2026-03-09T20:40:17.221 INFO:teuthology.orchestra.run.vm08.stdout:0 files removed 2026-03-09T20:40:17.244 DEBUG:teuthology.parallel:result is None 2026-03-09T20:40:17.244 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm03.local 2026-03-09T20:40:17.244 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm04.local 2026-03-09T20:40:17.244 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm08.local 2026-03-09T20:40:17.244 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T20:40:17.245 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T20:40:17.245 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T20:40:17.270 DEBUG:teuthology.orchestra.run.vm04:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T20:40:17.271 DEBUG:teuthology.orchestra.run.vm03:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T20:40:17.275 DEBUG:teuthology.orchestra.run.vm08:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T20:40:17.333 DEBUG:teuthology.parallel:result is None 2026-03-09T20:40:17.333 DEBUG:teuthology.parallel:result is None 2026-03-09T20:40:17.344 DEBUG:teuthology.parallel:result is None 2026-03-09T20:40:17.344 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-09T20:40:17.346 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-09T20:40:17.347 DEBUG:teuthology.orchestra.run.vm03:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T20:40:17.375 DEBUG:teuthology.orchestra.run.vm04:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T20:40:17.376 DEBUG:teuthology.orchestra.run.vm08:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T20:40:17.387 INFO:teuthology.orchestra.run.vm03.stderr:bash: line 1: ntpq: command not found 2026-03-09T20:40:17.389 INFO:teuthology.orchestra.run.vm04.stderr:bash: line 1: ntpq: command not found 2026-03-09T20:40:17.401 INFO:teuthology.orchestra.run.vm08.stderr:bash: line 1: ntpq: command not found 2026-03-09T20:40:17.454 INFO:teuthology.orchestra.run.vm03.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T20:40:17.454 INFO:teuthology.orchestra.run.vm03.stdout:=============================================================================== 2026-03-09T20:40:17.454 INFO:teuthology.orchestra.run.vm03.stdout:^+ router02.i-tk.de 2 6 277 54 -656us[-1017us] +/- 23ms 2026-03-09T20:40:17.454 INFO:teuthology.orchestra.run.vm03.stdout:^+ ntp0.vikings.net 2 6 377 56 +1628us[+1267us] +/- 56ms 2026-03-09T20:40:17.454 INFO:teuthology.orchestra.run.vm03.stdout:^+ static.179.181.75.5.clie> 3 6 377 56 -811us[-1172us] +/- 32ms 2026-03-09T20:40:17.454 INFO:teuthology.orchestra.run.vm03.stdout:^* cp.hypermediaa.de 2 6 77 51 +72us[ -289us] +/- 17ms 2026-03-09T20:40:17.455 INFO:teuthology.orchestra.run.vm08.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T20:40:17.455 INFO:teuthology.orchestra.run.vm08.stdout:=============================================================================== 2026-03-09T20:40:17.455 INFO:teuthology.orchestra.run.vm08.stdout:^* cp.hypermediaa.de 2 6 177 49 -162us[ -80us] +/- 17ms 2026-03-09T20:40:17.455 INFO:teuthology.orchestra.run.vm08.stdout:^+ router02.i-tk.de 2 6 277 53 +1177us[+1260us] +/- 25ms 2026-03-09T20:40:17.455 INFO:teuthology.orchestra.run.vm08.stdout:^+ ntp0.vikings.net 2 6 377 57 +1375us[+1458us] +/- 56ms 2026-03-09T20:40:17.455 INFO:teuthology.orchestra.run.vm08.stdout:^+ static.179.181.75.5.clie> 3 6 377 57 -1039us[ -956us] +/- 32ms 2026-03-09T20:40:17.455 INFO:teuthology.orchestra.run.vm04.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T20:40:17.455 INFO:teuthology.orchestra.run.vm04.stdout:=============================================================================== 2026-03-09T20:40:17.455 INFO:teuthology.orchestra.run.vm04.stdout:^+ router02.i-tk.de 2 6 277 51 +3175us[+3175us] +/- 28ms 2026-03-09T20:40:17.455 INFO:teuthology.orchestra.run.vm04.stdout:^+ ntp0.vikings.net 2 6 377 55 +1246us[+1349us] +/- 56ms 2026-03-09T20:40:17.455 INFO:teuthology.orchestra.run.vm04.stdout:^+ static.179.181.75.5.clie> 3 6 377 57 -1206us[-1103us] +/- 32ms 2026-03-09T20:40:17.455 INFO:teuthology.orchestra.run.vm04.stdout:^* cp.hypermediaa.de 2 6 377 54 -338us[ -235us] +/- 17ms 2026-03-09T20:40:17.455 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-09T20:40:17.457 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-09T20:40:17.458 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-09T20:40:17.460 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-09T20:40:17.462 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-09T20:40:17.464 INFO:teuthology.task.internal:Duration was 510.292640 seconds 2026-03-09T20:40:17.465 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-09T20:40:17.467 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-09T20:40:17.467 DEBUG:teuthology.orchestra.run.vm03:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T20:40:17.497 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T20:40:17.499 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T20:40:17.534 INFO:teuthology.orchestra.run.vm04.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T20:40:17.536 INFO:teuthology.orchestra.run.vm03.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T20:40:17.540 INFO:teuthology.orchestra.run.vm08.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T20:40:17.925 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-09T20:40:17.925 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm03.local 2026-03-09T20:40:17.925 DEBUG:teuthology.orchestra.run.vm03:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T20:40:17.988 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm04.local 2026-03-09T20:40:17.988 DEBUG:teuthology.orchestra.run.vm04:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T20:40:18.015 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm08.local 2026-03-09T20:40:18.015 DEBUG:teuthology.orchestra.run.vm08:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T20:40:18.039 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-09T20:40:18.039 DEBUG:teuthology.orchestra.run.vm03:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T20:40:18.041 DEBUG:teuthology.orchestra.run.vm04:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T20:40:18.057 DEBUG:teuthology.orchestra.run.vm08:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T20:40:18.468 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-09T20:40:18.468 DEBUG:teuthology.orchestra.run.vm03:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T20:40:18.469 DEBUG:teuthology.orchestra.run.vm04:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T20:40:18.470 DEBUG:teuthology.orchestra.run.vm08:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T20:40:18.492 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T20:40:18.492 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T20:40:18.493 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T20:40:18.493 INFO:teuthology.orchestra.run.vm08.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T20:40:18.493 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T20:40:18.493 INFO:teuthology.orchestra.run.vm03.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T20:40:18.493 INFO:teuthology.orchestra.run.vm08.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz/home/ubuntu/cephtest/archive/syslog/journalctl.log: 2026-03-09T20:40:18.493 INFO:teuthology.orchestra.run.vm03.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: gzip -5 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T20:40:18.493 INFO:teuthology.orchestra.run.vm03.stderr: --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T20:40:18.494 INFO:teuthology.orchestra.run.vm03.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz/home/ubuntu/cephtest/archive/syslog/journalctl.log: 2026-03-09T20:40:18.495 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T20:40:18.495 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T20:40:18.496 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T20:40:18.496 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T20:40:18.496 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-09T20:40:18.629 INFO:teuthology.orchestra.run.vm08.stderr: 98.3% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T20:40:18.631 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.3% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T20:40:18.635 INFO:teuthology.orchestra.run.vm03.stderr: 98.2% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T20:40:18.637 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-09T20:40:18.662 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-09T20:40:18.662 DEBUG:teuthology.orchestra.run.vm03:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T20:40:18.705 DEBUG:teuthology.orchestra.run.vm04:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T20:40:18.732 DEBUG:teuthology.orchestra.run.vm08:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T20:40:18.756 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-09T20:40:18.759 DEBUG:teuthology.orchestra.run.vm03:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T20:40:18.761 DEBUG:teuthology.orchestra.run.vm04:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T20:40:18.775 DEBUG:teuthology.orchestra.run.vm08:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T20:40:18.785 INFO:teuthology.orchestra.run.vm03.stdout:kernel.core_pattern = core 2026-03-09T20:40:18.802 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern = core 2026-03-09T20:40:18.821 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = core 2026-03-09T20:40:18.832 DEBUG:teuthology.orchestra.run.vm03:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T20:40:18.857 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:40:18.857 DEBUG:teuthology.orchestra.run.vm04:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T20:40:18.871 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:40:18.871 DEBUG:teuthology.orchestra.run.vm08:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T20:40:18.886 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T20:40:18.886 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-09T20:40:18.889 INFO:teuthology.task.internal:Transferring archived files... 2026-03-09T20:40:18.889 DEBUG:teuthology.misc:Transferring archived files from vm03:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/646/remote/vm03 2026-03-09T20:40:18.889 DEBUG:teuthology.orchestra.run.vm03:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T20:40:18.924 DEBUG:teuthology.misc:Transferring archived files from vm04:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/646/remote/vm04 2026-03-09T20:40:18.924 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T20:40:18.951 DEBUG:teuthology.misc:Transferring archived files from vm08:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/646/remote/vm08 2026-03-09T20:40:18.951 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T20:40:18.978 INFO:teuthology.task.internal:Removing archive directory... 2026-03-09T20:40:18.978 DEBUG:teuthology.orchestra.run.vm03:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T20:40:18.980 DEBUG:teuthology.orchestra.run.vm04:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T20:40:18.992 DEBUG:teuthology.orchestra.run.vm08:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T20:40:19.034 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-09T20:40:19.037 INFO:teuthology.task.internal:Not uploading archives. 2026-03-09T20:40:19.037 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-09T20:40:19.040 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-09T20:40:19.040 DEBUG:teuthology.orchestra.run.vm03:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T20:40:19.041 DEBUG:teuthology.orchestra.run.vm04:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T20:40:19.046 DEBUG:teuthology.orchestra.run.vm08:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T20:40:19.056 INFO:teuthology.orchestra.run.vm03.stdout: 8540550 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 20:40 /home/ubuntu/cephtest 2026-03-09T20:40:19.060 INFO:teuthology.orchestra.run.vm04.stdout: 8532144 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 20:40 /home/ubuntu/cephtest 2026-03-09T20:40:19.090 INFO:teuthology.orchestra.run.vm08.stdout: 8532144 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 20:40 /home/ubuntu/cephtest 2026-03-09T20:40:19.091 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-09T20:40:19.096 INFO:teuthology.run:Summary data: description: orch/cephadm/workunits/{0-distro/centos_9.stream agent/on mon_election/connectivity task/test_rgw_multisite} duration: 510.2926404476166 flavor: default owner: kyr success: true 2026-03-09T20:40:19.096 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-09T20:40:19.117 INFO:teuthology.run:pass