2026-03-09T17:37:18.473 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-09T17:37:18.478 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-09T17:37:18.495 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/590 branch: squid description: orch/cephadm/rbd_iscsi/{base/install cluster/{fixed-3 openstack} conf/{disable-pool-app} supported-container-hosts$/{centos_9.stream} workloads/cephadm_iscsi} email: null first_in_suite: false flavor: default job_id: '590' last_in_suite: false machine_type: vps name: kyr-2026-03-09_11:23:05-orch-squid-none-default-vps no_nested_subset: false openstack: - machine: cpus: 1 disk: 40 ram: 8000 volumes: count: 4 size: 30 os_type: centos os_version: 9.stream overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: global: mon warn on pool no app: false mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - MON_DOWN sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 - scontext=system_u:system_r:getty_t:s0 workunit: branch: tt-squid sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - mon.a - mgr.x - osd.0 - osd.1 - client.0 - ceph.iscsi.iscsi.a - - mon.b - osd.2 - osd.3 - osd.4 - client.1 - - mon.c - osd.5 - osd.6 - osd.7 - client.2 - ceph.iscsi.iscsi.b seed: 3443 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 targets: vm01.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBD1FrcgYgVu0Vv23lq7/QzfCxwH1E/MkQKLeWYhuV2A8OQURkI3XRA6dKYcrInpQGLq0arFd9/JCTm98l5ER9ok= vm04.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBO/D6/ODhqqS+HAAshMjJ9ggddK6BTMVagc7gi0ISDgpyaRdP/cTXKlgdBv1+d0SghGyxZ+jqwv43vDomA7Q9LY= vm05.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOapSGg/EsL3HCzTkSixBTOGnali/unzDbgyLJ2aEVD8gj/7LQYlN4diZN2XxmwG67ODKchlAh95RdYHvNsJnts= tasks: - cephadm: null - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - install: extra_system_packages: deb: - open-iscsi - multipath-tools rpm: - iscsi-initiator-utils - device-mapper-multipath - pexec: all: - sudo dnf remove nvme-cli -y - sudo dnf install nvmetcli nvme-cli -y - ceph_iscsi_client: clients: - client.1 - cram: clients: client.0: - src/test/cli-integration/rbd/gwcli_create.t client.1: - src/test/cli-integration/rbd/iscsi_client.t client.2: - src/test/cli-integration/rbd/gwcli_delete.t parallel: false - cram: clients: client.0: - src/test/cli-integration/rbd/rest_api_create.t client.1: - src/test/cli-integration/rbd/iscsi_client.t client.2: - src/test/cli-integration/rbd/rest_api_delete.t parallel: false teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-09_11:23:05 tube: vps use_shaman: true user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.611473 2026-03-09T17:37:18.496 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa; will attempt to use it 2026-03-09T17:37:18.496 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa/tasks 2026-03-09T17:37:18.496 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-09T17:37:18.496 INFO:teuthology.task.internal:Checking packages... 2026-03-09T17:37:18.496 INFO:teuthology.task.internal:Checking packages for os_type 'centos', flavor 'default' and ceph hash 'e911bdebe5c8faa3800735d1568fcdca65db60df' 2026-03-09T17:37:18.496 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-09T17:37:18.496 INFO:teuthology.packaging:ref: None 2026-03-09T17:37:18.496 INFO:teuthology.packaging:tag: None 2026-03-09T17:37:18.496 INFO:teuthology.packaging:branch: squid 2026-03-09T17:37:18.496 INFO:teuthology.packaging:sha1: e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T17:37:18.496 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&ref=squid 2026-03-09T17:37:19.270 INFO:teuthology.task.internal:Found packages for ceph version 19.2.3-678.ge911bdeb 2026-03-09T17:37:19.271 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-09T17:37:19.272 INFO:teuthology.task.internal:no buildpackages task found 2026-03-09T17:37:19.272 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-09T17:37:19.272 INFO:teuthology.task.internal:Saving configuration 2026-03-09T17:37:19.277 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-09T17:37:19.278 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-09T17:37:19.284 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm01.local', 'description': '/archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/590', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-09 17:34:45.064577', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:01', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBD1FrcgYgVu0Vv23lq7/QzfCxwH1E/MkQKLeWYhuV2A8OQURkI3XRA6dKYcrInpQGLq0arFd9/JCTm98l5ER9ok='} 2026-03-09T17:37:19.288 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm04.local', 'description': '/archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/590', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-09 17:34:45.065248', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:04', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBO/D6/ODhqqS+HAAshMjJ9ggddK6BTMVagc7gi0ISDgpyaRdP/cTXKlgdBv1+d0SghGyxZ+jqwv43vDomA7Q9LY='} 2026-03-09T17:37:19.293 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm05.local', 'description': '/archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/590', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-09 17:34:45.065018', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:05', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOapSGg/EsL3HCzTkSixBTOGnali/unzDbgyLJ2aEVD8gj/7LQYlN4diZN2XxmwG67ODKchlAh95RdYHvNsJnts='} 2026-03-09T17:37:19.293 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-09T17:37:19.294 INFO:teuthology.task.internal:roles: ubuntu@vm01.local - ['host.a', 'mon.a', 'mgr.x', 'osd.0', 'osd.1', 'client.0', 'ceph.iscsi.iscsi.a'] 2026-03-09T17:37:19.294 INFO:teuthology.task.internal:roles: ubuntu@vm04.local - ['mon.b', 'osd.2', 'osd.3', 'osd.4', 'client.1'] 2026-03-09T17:37:19.294 INFO:teuthology.task.internal:roles: ubuntu@vm05.local - ['mon.c', 'osd.5', 'osd.6', 'osd.7', 'client.2', 'ceph.iscsi.iscsi.b'] 2026-03-09T17:37:19.294 INFO:teuthology.run_tasks:Running task console_log... 2026-03-09T17:37:19.299 DEBUG:teuthology.task.console_log:vm01 does not support IPMI; excluding 2026-03-09T17:37:19.303 DEBUG:teuthology.task.console_log:vm04 does not support IPMI; excluding 2026-03-09T17:37:19.307 DEBUG:teuthology.task.console_log:vm05 does not support IPMI; excluding 2026-03-09T17:37:19.307 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7f4b59972170>, signals=[15]) 2026-03-09T17:37:19.307 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-09T17:37:19.308 INFO:teuthology.task.internal:Opening connections... 2026-03-09T17:37:19.308 DEBUG:teuthology.task.internal:connecting to ubuntu@vm01.local 2026-03-09T17:37:19.309 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T17:37:19.370 DEBUG:teuthology.task.internal:connecting to ubuntu@vm04.local 2026-03-09T17:37:19.370 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm04.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T17:37:19.430 DEBUG:teuthology.task.internal:connecting to ubuntu@vm05.local 2026-03-09T17:37:19.431 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm05.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T17:37:19.492 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-09T17:37:19.493 DEBUG:teuthology.orchestra.run.vm01:> uname -m 2026-03-09T17:37:19.510 INFO:teuthology.orchestra.run.vm01.stdout:x86_64 2026-03-09T17:37:19.510 DEBUG:teuthology.orchestra.run.vm01:> cat /etc/os-release 2026-03-09T17:37:19.567 INFO:teuthology.orchestra.run.vm01.stdout:NAME="CentOS Stream" 2026-03-09T17:37:19.567 INFO:teuthology.orchestra.run.vm01.stdout:VERSION="9" 2026-03-09T17:37:19.567 INFO:teuthology.orchestra.run.vm01.stdout:ID="centos" 2026-03-09T17:37:19.567 INFO:teuthology.orchestra.run.vm01.stdout:ID_LIKE="rhel fedora" 2026-03-09T17:37:19.567 INFO:teuthology.orchestra.run.vm01.stdout:VERSION_ID="9" 2026-03-09T17:37:19.567 INFO:teuthology.orchestra.run.vm01.stdout:PLATFORM_ID="platform:el9" 2026-03-09T17:37:19.567 INFO:teuthology.orchestra.run.vm01.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T17:37:19.567 INFO:teuthology.orchestra.run.vm01.stdout:ANSI_COLOR="0;31" 2026-03-09T17:37:19.567 INFO:teuthology.orchestra.run.vm01.stdout:LOGO="fedora-logo-icon" 2026-03-09T17:37:19.567 INFO:teuthology.orchestra.run.vm01.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T17:37:19.567 INFO:teuthology.orchestra.run.vm01.stdout:HOME_URL="https://centos.org/" 2026-03-09T17:37:19.567 INFO:teuthology.orchestra.run.vm01.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T17:37:19.567 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T17:37:19.567 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T17:37:19.567 INFO:teuthology.lock.ops:Updating vm01.local on lock server 2026-03-09T17:37:19.571 DEBUG:teuthology.orchestra.run.vm04:> uname -m 2026-03-09T17:37:19.589 INFO:teuthology.orchestra.run.vm04.stdout:x86_64 2026-03-09T17:37:19.590 DEBUG:teuthology.orchestra.run.vm04:> cat /etc/os-release 2026-03-09T17:37:19.646 INFO:teuthology.orchestra.run.vm04.stdout:NAME="CentOS Stream" 2026-03-09T17:37:19.646 INFO:teuthology.orchestra.run.vm04.stdout:VERSION="9" 2026-03-09T17:37:19.646 INFO:teuthology.orchestra.run.vm04.stdout:ID="centos" 2026-03-09T17:37:19.646 INFO:teuthology.orchestra.run.vm04.stdout:ID_LIKE="rhel fedora" 2026-03-09T17:37:19.646 INFO:teuthology.orchestra.run.vm04.stdout:VERSION_ID="9" 2026-03-09T17:37:19.646 INFO:teuthology.orchestra.run.vm04.stdout:PLATFORM_ID="platform:el9" 2026-03-09T17:37:19.646 INFO:teuthology.orchestra.run.vm04.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T17:37:19.647 INFO:teuthology.orchestra.run.vm04.stdout:ANSI_COLOR="0;31" 2026-03-09T17:37:19.647 INFO:teuthology.orchestra.run.vm04.stdout:LOGO="fedora-logo-icon" 2026-03-09T17:37:19.647 INFO:teuthology.orchestra.run.vm04.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T17:37:19.647 INFO:teuthology.orchestra.run.vm04.stdout:HOME_URL="https://centos.org/" 2026-03-09T17:37:19.647 INFO:teuthology.orchestra.run.vm04.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T17:37:19.647 INFO:teuthology.orchestra.run.vm04.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T17:37:19.647 INFO:teuthology.orchestra.run.vm04.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T17:37:19.647 INFO:teuthology.lock.ops:Updating vm04.local on lock server 2026-03-09T17:37:19.651 DEBUG:teuthology.orchestra.run.vm05:> uname -m 2026-03-09T17:37:19.665 INFO:teuthology.orchestra.run.vm05.stdout:x86_64 2026-03-09T17:37:19.665 DEBUG:teuthology.orchestra.run.vm05:> cat /etc/os-release 2026-03-09T17:37:19.721 INFO:teuthology.orchestra.run.vm05.stdout:NAME="CentOS Stream" 2026-03-09T17:37:19.721 INFO:teuthology.orchestra.run.vm05.stdout:VERSION="9" 2026-03-09T17:37:19.721 INFO:teuthology.orchestra.run.vm05.stdout:ID="centos" 2026-03-09T17:37:19.721 INFO:teuthology.orchestra.run.vm05.stdout:ID_LIKE="rhel fedora" 2026-03-09T17:37:19.721 INFO:teuthology.orchestra.run.vm05.stdout:VERSION_ID="9" 2026-03-09T17:37:19.721 INFO:teuthology.orchestra.run.vm05.stdout:PLATFORM_ID="platform:el9" 2026-03-09T17:37:19.721 INFO:teuthology.orchestra.run.vm05.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T17:37:19.721 INFO:teuthology.orchestra.run.vm05.stdout:ANSI_COLOR="0;31" 2026-03-09T17:37:19.721 INFO:teuthology.orchestra.run.vm05.stdout:LOGO="fedora-logo-icon" 2026-03-09T17:37:19.721 INFO:teuthology.orchestra.run.vm05.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T17:37:19.721 INFO:teuthology.orchestra.run.vm05.stdout:HOME_URL="https://centos.org/" 2026-03-09T17:37:19.721 INFO:teuthology.orchestra.run.vm05.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T17:37:19.721 INFO:teuthology.orchestra.run.vm05.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T17:37:19.721 INFO:teuthology.orchestra.run.vm05.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T17:37:19.721 INFO:teuthology.lock.ops:Updating vm05.local on lock server 2026-03-09T17:37:19.726 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-09T17:37:19.728 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-09T17:37:19.729 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-09T17:37:19.729 DEBUG:teuthology.orchestra.run.vm01:> test '!' -e /home/ubuntu/cephtest 2026-03-09T17:37:19.730 DEBUG:teuthology.orchestra.run.vm04:> test '!' -e /home/ubuntu/cephtest 2026-03-09T17:37:19.732 DEBUG:teuthology.orchestra.run.vm05:> test '!' -e /home/ubuntu/cephtest 2026-03-09T17:37:19.775 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-09T17:37:19.777 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-09T17:37:19.777 DEBUG:teuthology.orchestra.run.vm01:> test -z $(ls -A /var/lib/ceph) 2026-03-09T17:37:19.787 DEBUG:teuthology.orchestra.run.vm04:> test -z $(ls -A /var/lib/ceph) 2026-03-09T17:37:19.791 DEBUG:teuthology.orchestra.run.vm05:> test -z $(ls -A /var/lib/ceph) 2026-03-09T17:37:19.800 INFO:teuthology.orchestra.run.vm01.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T17:37:19.809 INFO:teuthology.orchestra.run.vm04.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T17:37:19.831 INFO:teuthology.orchestra.run.vm05.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T17:37:19.832 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-09T17:37:19.840 DEBUG:teuthology.orchestra.run.vm01:> test -e /ceph-qa-ready 2026-03-09T17:37:19.856 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:37:20.052 DEBUG:teuthology.orchestra.run.vm04:> test -e /ceph-qa-ready 2026-03-09T17:37:20.070 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:37:20.273 DEBUG:teuthology.orchestra.run.vm05:> test -e /ceph-qa-ready 2026-03-09T17:37:20.289 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:37:20.476 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-09T17:37:20.477 INFO:teuthology.task.internal:Creating test directory... 2026-03-09T17:37:20.477 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T17:37:20.479 DEBUG:teuthology.orchestra.run.vm04:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T17:37:20.481 DEBUG:teuthology.orchestra.run.vm05:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T17:37:20.500 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-09T17:37:20.502 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-09T17:37:20.503 INFO:teuthology.task.internal:Creating archive directory... 2026-03-09T17:37:20.503 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T17:37:20.537 DEBUG:teuthology.orchestra.run.vm04:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T17:37:20.543 DEBUG:teuthology.orchestra.run.vm05:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T17:37:20.560 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-09T17:37:20.562 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-09T17:37:20.562 DEBUG:teuthology.orchestra.run.vm01:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T17:37:20.604 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:37:20.604 DEBUG:teuthology.orchestra.run.vm04:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T17:37:20.618 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:37:20.618 DEBUG:teuthology.orchestra.run.vm05:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T17:37:20.632 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:37:20.632 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T17:37:20.646 DEBUG:teuthology.orchestra.run.vm04:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T17:37:20.660 DEBUG:teuthology.orchestra.run.vm05:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T17:37:20.671 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T17:37:20.682 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T17:37:20.686 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T17:37:20.695 INFO:teuthology.orchestra.run.vm05.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T17:37:20.696 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T17:37:20.704 INFO:teuthology.orchestra.run.vm05.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T17:37:20.706 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-09T17:37:20.707 INFO:teuthology.task.internal:Configuring sudo... 2026-03-09T17:37:20.707 DEBUG:teuthology.orchestra.run.vm01:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T17:37:20.725 DEBUG:teuthology.orchestra.run.vm04:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T17:37:20.738 DEBUG:teuthology.orchestra.run.vm05:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T17:37:20.772 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-09T17:37:20.774 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-09T17:37:20.774 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T17:37:20.792 DEBUG:teuthology.orchestra.run.vm04:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T17:37:20.803 DEBUG:teuthology.orchestra.run.vm05:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T17:37:20.829 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T17:37:20.870 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T17:37:20.927 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:37:20.927 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T17:37:20.985 DEBUG:teuthology.orchestra.run.vm04:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T17:37:21.011 DEBUG:teuthology.orchestra.run.vm04:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T17:37:21.071 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:37:21.071 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T17:37:21.135 DEBUG:teuthology.orchestra.run.vm05:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T17:37:21.157 DEBUG:teuthology.orchestra.run.vm05:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T17:37:21.214 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-09T17:37:21.215 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T17:37:21.273 DEBUG:teuthology.orchestra.run.vm01:> sudo service rsyslog restart 2026-03-09T17:37:21.275 DEBUG:teuthology.orchestra.run.vm04:> sudo service rsyslog restart 2026-03-09T17:37:21.276 DEBUG:teuthology.orchestra.run.vm05:> sudo service rsyslog restart 2026-03-09T17:37:21.299 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T17:37:21.305 INFO:teuthology.orchestra.run.vm04.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T17:37:21.339 INFO:teuthology.orchestra.run.vm05.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T17:37:21.759 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-09T17:37:21.761 INFO:teuthology.task.internal:Starting timer... 2026-03-09T17:37:21.761 INFO:teuthology.run_tasks:Running task pcp... 2026-03-09T17:37:21.763 INFO:teuthology.run_tasks:Running task selinux... 2026-03-09T17:37:21.766 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0', 'scontext=system_u:system_r:getty_t:s0']} 2026-03-09T17:37:21.766 INFO:teuthology.task.selinux:Excluding vm01: VMs are not yet supported 2026-03-09T17:37:21.766 INFO:teuthology.task.selinux:Excluding vm04: VMs are not yet supported 2026-03-09T17:37:21.766 INFO:teuthology.task.selinux:Excluding vm05: VMs are not yet supported 2026-03-09T17:37:21.766 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-09T17:37:21.766 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-09T17:37:21.766 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-09T17:37:21.766 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-09T17:37:21.767 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-09T17:37:21.768 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-09T17:37:21.770 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-09T17:37:22.353 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-09T17:37:22.359 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-09T17:37:22.359 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventorykfu4wu61 --limit vm01.local,vm04.local,vm05.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-09T17:40:03.111 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm01.local'), Remote(name='ubuntu@vm04.local'), Remote(name='ubuntu@vm05.local')] 2026-03-09T17:40:03.111 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm01.local' 2026-03-09T17:40:03.111 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T17:40:03.173 DEBUG:teuthology.orchestra.run.vm01:> true 2026-03-09T17:40:03.253 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm01.local' 2026-03-09T17:40:03.253 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm04.local' 2026-03-09T17:40:03.253 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm04.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T17:40:03.318 DEBUG:teuthology.orchestra.run.vm04:> true 2026-03-09T17:40:03.400 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm04.local' 2026-03-09T17:40:03.400 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm05.local' 2026-03-09T17:40:03.400 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm05.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T17:40:03.465 DEBUG:teuthology.orchestra.run.vm05:> true 2026-03-09T17:40:03.548 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm05.local' 2026-03-09T17:40:03.548 INFO:teuthology.run_tasks:Running task clock... 2026-03-09T17:40:03.551 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-09T17:40:03.551 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T17:40:03.551 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T17:40:03.553 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T17:40:03.553 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T17:40:03.556 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T17:40:03.556 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T17:40:03.592 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T17:40:03.595 INFO:teuthology.orchestra.run.vm04.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T17:40:03.609 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T17:40:03.614 INFO:teuthology.orchestra.run.vm04.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T17:40:03.625 INFO:teuthology.orchestra.run.vm05.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T17:40:03.641 INFO:teuthology.orchestra.run.vm05.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T17:40:03.644 INFO:teuthology.orchestra.run.vm01.stderr:sudo: ntpd: command not found 2026-03-09T17:40:03.652 INFO:teuthology.orchestra.run.vm04.stderr:sudo: ntpd: command not found 2026-03-09T17:40:03.660 INFO:teuthology.orchestra.run.vm01.stdout:506 Cannot talk to daemon 2026-03-09T17:40:03.666 INFO:teuthology.orchestra.run.vm04.stdout:506 Cannot talk to daemon 2026-03-09T17:40:03.670 INFO:teuthology.orchestra.run.vm05.stderr:sudo: ntpd: command not found 2026-03-09T17:40:03.678 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T17:40:03.685 INFO:teuthology.orchestra.run.vm04.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T17:40:03.686 INFO:teuthology.orchestra.run.vm05.stdout:506 Cannot talk to daemon 2026-03-09T17:40:03.693 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T17:40:03.702 INFO:teuthology.orchestra.run.vm05.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T17:40:03.703 INFO:teuthology.orchestra.run.vm04.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T17:40:03.718 INFO:teuthology.orchestra.run.vm05.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T17:40:03.755 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-09T17:40:03.755 INFO:teuthology.orchestra.run.vm04.stderr:bash: line 1: ntpq: command not found 2026-03-09T17:40:03.757 INFO:teuthology.orchestra.run.vm04.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T17:40:03.758 INFO:teuthology.orchestra.run.vm04.stdout:=============================================================================== 2026-03-09T17:40:03.758 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T17:40:03.758 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-09T17:40:03.765 INFO:teuthology.orchestra.run.vm05.stderr:bash: line 1: ntpq: command not found 2026-03-09T17:40:03.767 INFO:teuthology.orchestra.run.vm05.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T17:40:03.767 INFO:teuthology.orchestra.run.vm05.stdout:=============================================================================== 2026-03-09T17:40:03.768 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-09T17:40:03.817 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon warn on pool no app': False}, 'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'MON_DOWN'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-09T17:40:03.817 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T17:40:03.817 INFO:tasks.cephadm:Cluster fsid is 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:40:03.817 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-09T17:40:03.817 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.101', 'mon.b': '192.168.123.104', 'mon.c': '192.168.123.105'} 2026-03-09T17:40:03.817 INFO:tasks.cephadm:First mon is mon.a on vm01 2026-03-09T17:40:03.817 INFO:tasks.cephadm:First mgr is x 2026-03-09T17:40:03.817 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-09T17:40:03.817 DEBUG:teuthology.orchestra.run.vm01:> sudo hostname $(hostname -s) 2026-03-09T17:40:03.854 DEBUG:teuthology.orchestra.run.vm04:> sudo hostname $(hostname -s) 2026-03-09T17:40:03.889 DEBUG:teuthology.orchestra.run.vm05:> sudo hostname $(hostname -s) 2026-03-09T17:40:03.925 INFO:tasks.cephadm:Downloading "compiled" cephadm from cachra 2026-03-09T17:40:03.925 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T17:40:04.599 INFO:tasks.cephadm:builder_project result: [{'url': 'https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'chacra_url': 'https://3.chacra.ceph.com/repos/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'ref': 'squid', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'distro': 'centos', 'distro_version': '9', 'distro_codename': None, 'modified': '2026-02-25 18:55:15.146628', 'status': 'ready', 'flavor': 'default', 'project': 'ceph', 'archs': ['source', 'x86_64'], 'extra': {'version': '19.2.3-678-ge911bdeb', 'package_manager_version': '19.2.3-678.ge911bdeb', 'build_url': 'https://jenkins.ceph.com/job/ceph-dev-pipeline/3275/', 'root_build_cause': '', 'node_name': '10.20.192.26+soko16', 'job_name': 'ceph-dev-pipeline'}}] 2026-03-09T17:40:05.274 INFO:tasks.util.chacra:got chacra host 3.chacra.ceph.com, ref squid, sha1 e911bdebe5c8faa3800735d1568fcdca65db60df from https://shaman.ceph.com/api/search/?project=ceph&distros=centos%2F9%2Fx86_64&flavor=default&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T17:40:05.275 INFO:tasks.cephadm:Discovered cachra url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-09T17:40:05.275 INFO:tasks.cephadm:Downloading cephadm from url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-09T17:40:05.275 DEBUG:teuthology.orchestra.run.vm01:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T17:40:06.763 INFO:teuthology.orchestra.run.vm01.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 17:40 /home/ubuntu/cephtest/cephadm 2026-03-09T17:40:06.763 DEBUG:teuthology.orchestra.run.vm04:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T17:40:08.210 INFO:teuthology.orchestra.run.vm04.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 17:40 /home/ubuntu/cephtest/cephadm 2026-03-09T17:40:08.210 DEBUG:teuthology.orchestra.run.vm05:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T17:40:09.584 INFO:teuthology.orchestra.run.vm05.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 17:40 /home/ubuntu/cephtest/cephadm 2026-03-09T17:40:09.584 DEBUG:teuthology.orchestra.run.vm01:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T17:40:09.605 DEBUG:teuthology.orchestra.run.vm04:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T17:40:09.629 DEBUG:teuthology.orchestra.run.vm05:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T17:40:09.656 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df on all hosts... 2026-03-09T17:40:09.656 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T17:40:09.659 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T17:40:09.674 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T17:40:09.868 INFO:teuthology.orchestra.run.vm01.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T17:40:09.877 INFO:teuthology.orchestra.run.vm04.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T17:40:09.902 INFO:teuthology.orchestra.run.vm05.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T17:40:50.319 INFO:teuthology.orchestra.run.vm04.stdout:{ 2026-03-09T17:40:50.319 INFO:teuthology.orchestra.run.vm04.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T17:40:50.319 INFO:teuthology.orchestra.run.vm04.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T17:40:50.319 INFO:teuthology.orchestra.run.vm04.stdout: "repo_digests": [ 2026-03-09T17:40:50.319 INFO:teuthology.orchestra.run.vm04.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T17:40:50.319 INFO:teuthology.orchestra.run.vm04.stdout: ] 2026-03-09T17:40:50.319 INFO:teuthology.orchestra.run.vm04.stdout:} 2026-03-09T17:41:01.322 INFO:teuthology.orchestra.run.vm01.stdout:{ 2026-03-09T17:41:01.322 INFO:teuthology.orchestra.run.vm01.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T17:41:01.322 INFO:teuthology.orchestra.run.vm01.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T17:41:01.322 INFO:teuthology.orchestra.run.vm01.stdout: "repo_digests": [ 2026-03-09T17:41:01.322 INFO:teuthology.orchestra.run.vm01.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T17:41:01.322 INFO:teuthology.orchestra.run.vm01.stdout: ] 2026-03-09T17:41:01.322 INFO:teuthology.orchestra.run.vm01.stdout:} 2026-03-09T17:41:04.627 INFO:teuthology.orchestra.run.vm05.stdout:{ 2026-03-09T17:41:04.628 INFO:teuthology.orchestra.run.vm05.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T17:41:04.628 INFO:teuthology.orchestra.run.vm05.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T17:41:04.628 INFO:teuthology.orchestra.run.vm05.stdout: "repo_digests": [ 2026-03-09T17:41:04.628 INFO:teuthology.orchestra.run.vm05.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T17:41:04.628 INFO:teuthology.orchestra.run.vm05.stdout: ] 2026-03-09T17:41:04.628 INFO:teuthology.orchestra.run.vm05.stdout:} 2026-03-09T17:41:04.646 DEBUG:teuthology.orchestra.run.vm01:> sudo mkdir -p /etc/ceph 2026-03-09T17:41:04.682 DEBUG:teuthology.orchestra.run.vm04:> sudo mkdir -p /etc/ceph 2026-03-09T17:41:04.711 DEBUG:teuthology.orchestra.run.vm05:> sudo mkdir -p /etc/ceph 2026-03-09T17:41:04.738 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod 777 /etc/ceph 2026-03-09T17:41:04.769 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod 777 /etc/ceph 2026-03-09T17:41:04.794 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod 777 /etc/ceph 2026-03-09T17:41:04.819 INFO:tasks.cephadm:Writing seed config... 2026-03-09T17:41:04.819 INFO:tasks.cephadm: override: [global] mon warn on pool no app = False 2026-03-09T17:41:04.819 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-09T17:41:04.819 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-09T17:41:04.819 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-09T17:41:04.819 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-09T17:41:04.819 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-09T17:41:04.819 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-09T17:41:04.819 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-09T17:41:04.819 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-09T17:41:04.819 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:41:04.819 DEBUG:teuthology.orchestra.run.vm01:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-09T17:41:04.847 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 01455850-1bdf-11f1-910a-9936d43313cc mon warn on pool no app = False [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-09T17:41:04.847 DEBUG:teuthology.orchestra.run.vm01:mon.a> sudo journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.a.service 2026-03-09T17:41:04.892 DEBUG:teuthology.orchestra.run.vm01:mgr.x> sudo journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@mgr.x.service 2026-03-09T17:41:04.934 INFO:tasks.cephadm:Bootstrapping... 2026-03-09T17:41:04.934 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df -v bootstrap --fsid 01455850-1bdf-11f1-910a-9936d43313cc --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id x --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.101 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-09T17:41:05.100 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-09T17:41:05.100 INFO:teuthology.orchestra.run.vm01.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df', '-v', 'bootstrap', '--fsid', '01455850-1bdf-11f1-910a-9936d43313cc', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'x', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.101', '--skip-admin-label'] 2026-03-09T17:41:05.101 INFO:teuthology.orchestra.run.vm01.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-09T17:41:05.101 INFO:teuthology.orchestra.run.vm01.stdout:Verifying podman|docker is present... 2026-03-09T17:41:05.128 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 5.8.0 2026-03-09T17:41:05.128 INFO:teuthology.orchestra.run.vm01.stdout:Verifying lvm2 is present... 2026-03-09T17:41:05.128 INFO:teuthology.orchestra.run.vm01.stdout:Verifying time synchronization is in place... 2026-03-09T17:41:05.138 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-09T17:41:05.139 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-09T17:41:05.145 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-09T17:41:05.145 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-09T17:41:05.151 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout enabled 2026-03-09T17:41:05.156 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout active 2026-03-09T17:41:05.156 INFO:teuthology.orchestra.run.vm01.stdout:Unit chronyd.service is enabled and running 2026-03-09T17:41:05.156 INFO:teuthology.orchestra.run.vm01.stdout:Repeating the final host check... 2026-03-09T17:41:05.176 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 5.8.0 2026-03-09T17:41:05.176 INFO:teuthology.orchestra.run.vm01.stdout:podman (/bin/podman) version 5.8.0 is present 2026-03-09T17:41:05.176 INFO:teuthology.orchestra.run.vm01.stdout:systemctl is present 2026-03-09T17:41:05.176 INFO:teuthology.orchestra.run.vm01.stdout:lvcreate is present 2026-03-09T17:41:05.182 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-09T17:41:05.182 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-09T17:41:05.188 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-09T17:41:05.188 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-09T17:41:05.194 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout enabled 2026-03-09T17:41:05.200 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout active 2026-03-09T17:41:05.200 INFO:teuthology.orchestra.run.vm01.stdout:Unit chronyd.service is enabled and running 2026-03-09T17:41:05.200 INFO:teuthology.orchestra.run.vm01.stdout:Host looks OK 2026-03-09T17:41:05.200 INFO:teuthology.orchestra.run.vm01.stdout:Cluster fsid: 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:05.200 INFO:teuthology.orchestra.run.vm01.stdout:Acquiring lock 140073970202320 on /run/cephadm/01455850-1bdf-11f1-910a-9936d43313cc.lock 2026-03-09T17:41:05.200 INFO:teuthology.orchestra.run.vm01.stdout:Lock 140073970202320 acquired on /run/cephadm/01455850-1bdf-11f1-910a-9936d43313cc.lock 2026-03-09T17:41:05.200 INFO:teuthology.orchestra.run.vm01.stdout:Verifying IP 192.168.123.101 port 3300 ... 2026-03-09T17:41:05.200 INFO:teuthology.orchestra.run.vm01.stdout:Verifying IP 192.168.123.101 port 6789 ... 2026-03-09T17:41:05.201 INFO:teuthology.orchestra.run.vm01.stdout:Base mon IP(s) is [192.168.123.101:3300, 192.168.123.101:6789], mon addrv is [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-09T17:41:05.204 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.101 metric 100 2026-03-09T17:41:05.204 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.101 metric 100 2026-03-09T17:41:05.207 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-09T17:41:05.207 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-09T17:41:05.209 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-09T17:41:05.209 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-09T17:41:05.209 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-09T17:41:05.209 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 2: eth0: mtu 1500 state UP qlen 1000 2026-03-09T17:41:05.209 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout inet6 fe80::5055:ff:fe00:1/64 scope link noprefixroute 2026-03-09T17:41:05.209 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-09T17:41:05.209 INFO:teuthology.orchestra.run.vm01.stdout:Mon IP `192.168.123.101` is in CIDR network `192.168.123.0/24` 2026-03-09T17:41:05.209 INFO:teuthology.orchestra.run.vm01.stdout:Mon IP `192.168.123.101` is in CIDR network `192.168.123.0/24` 2026-03-09T17:41:05.210 INFO:teuthology.orchestra.run.vm01.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24'] 2026-03-09T17:41:05.210 INFO:teuthology.orchestra.run.vm01.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-09T17:41:05.210 INFO:teuthology.orchestra.run.vm01.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T17:41:06.762 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-09T17:41:06.762 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Trying to pull quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T17:41:06.762 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Getting image source signatures 2026-03-09T17:41:06.762 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Copying blob sha256:1752b8d01aa0dd33bbe0ab24e8316174c94fbdcd5d26252e2680bba0624747a7 2026-03-09T17:41:06.762 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Copying blob sha256:8e380faede39ebd4286247457b408d979ab568aafd8389c42ec304b8cfba4e92 2026-03-09T17:41:06.762 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Copying config sha256:654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-09T17:41:06.762 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Writing manifest to image destination 2026-03-09T17:41:06.903 INFO:teuthology.orchestra.run.vm01.stdout:ceph: stdout ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-09T17:41:06.903 INFO:teuthology.orchestra.run.vm01.stdout:Ceph version: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-09T17:41:06.903 INFO:teuthology.orchestra.run.vm01.stdout:Extracting ceph user uid/gid from container image... 2026-03-09T17:41:06.999 INFO:teuthology.orchestra.run.vm01.stdout:stat: stdout 167 167 2026-03-09T17:41:06.999 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial keys... 2026-03-09T17:41:07.111 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQAzBq9pFrb+BBAAP0DN88adAZhjD8pLy3qB8Q== 2026-03-09T17:41:07.231 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQAzBq9pEVU6CxAA1VXXG3Nu/cYvp1p3jHBVUQ== 2026-03-09T17:41:07.351 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQAzBq9pWmoMEhAAOoky4vuOSBecOC+sND3ZLw== 2026-03-09T17:41:07.352 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial monmap... 2026-03-09T17:41:07.455 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-09T17:41:07.455 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-09T17:41:07.455 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:07.455 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-09T17:41:07.455 INFO:teuthology.orchestra.run.vm01.stdout:monmaptool for a [v2:192.168.123.101:3300,v1:192.168.123.101:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-09T17:41:07.455 INFO:teuthology.orchestra.run.vm01.stdout:setting min_mon_release = quincy 2026-03-09T17:41:07.455 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: set fsid to 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:07.455 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-09T17:41:07.455 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:41:07.455 INFO:teuthology.orchestra.run.vm01.stdout:Creating mon... 2026-03-09T17:41:07.608 INFO:teuthology.orchestra.run.vm01.stdout:create mon.a on 2026-03-09T17:41:07.932 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-09T17:41:08.085 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-01455850-1bdf-11f1-910a-9936d43313cc.target → /etc/systemd/system/ceph-01455850-1bdf-11f1-910a-9936d43313cc.target. 2026-03-09T17:41:08.085 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-01455850-1bdf-11f1-910a-9936d43313cc.target → /etc/systemd/system/ceph-01455850-1bdf-11f1-910a-9936d43313cc.target. 2026-03-09T17:41:08.256 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.a 2026-03-09T17:41:08.257 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to reset failed state of unit ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.a.service: Unit ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.a.service not loaded. 2026-03-09T17:41:08.425 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-01455850-1bdf-11f1-910a-9936d43313cc.target.wants/ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.a.service → /etc/systemd/system/ceph-01455850-1bdf-11f1-910a-9936d43313cc@.service. 2026-03-09T17:41:08.597 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:08 vm01 podman[52501]: 2026-03-09 17:41:08.589391227 +0000 UTC m=+0.057746993 container start 3d6814385085dd8644a6824b89f9dbdb64d0ee3d58b5bd45643bbefeaeaafd97 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-a, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, OSD_FLAVOR=default, ceph=True, org.label-schema.license=GPLv2) 2026-03-09T17:41:08.607 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-09T17:41:08.607 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to enable service . firewalld.service is not available 2026-03-09T17:41:08.607 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mon to start... 2026-03-09T17:41:08.607 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mon... 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout cluster: 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout id: 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout services: 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.130437s) 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout data: 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout pgs: 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:mon is available 2026-03-09T17:41:08.813 INFO:teuthology.orchestra.run.vm01.stdout:Assimilating anything we can from ceph.conf... 2026-03-09T17:41:08.850 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:08 vm01 bash[52501]: 3d6814385085dd8644a6824b89f9dbdb64d0ee3d58b5bd45643bbefeaeaafd97 2026-03-09T17:41:08.850 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:08 vm01 podman[52501]: 2026-03-09 17:41:08.542213318 +0000 UTC m=+0.010569084 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T17:41:08.850 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:08 vm01 systemd[1]: Started Ceph mon.a for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:41:08.850 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:08 vm01 ceph-mon[52515]: mkfs 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:08.850 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:08 vm01 ceph-mon[52515]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [global] 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout fsid = 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [osd] 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-09T17:41:09.016 INFO:teuthology.orchestra.run.vm01.stdout:Generating new minimal ceph.conf... 2026-03-09T17:41:09.200 INFO:teuthology.orchestra.run.vm01.stdout:Restarting the monitor... 2026-03-09T17:41:09.349 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 systemd[1]: Stopping Ceph mon.a for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:41:09.349 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-a[52511]: 2026-03-09T17:41:09.278+0000 7fc2f636c640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T17:41:09.349 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-a[52511]: 2026-03-09T17:41:09.278+0000 7fc2f636c640 -1 mon.a@0(leader) e1 *** Got Signal Terminated *** 2026-03-09T17:41:09.600 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 podman[52712]: 2026-03-09 17:41:09.535792356 +0000 UTC m=+0.271613124 container died 3d6814385085dd8644a6824b89f9dbdb64d0ee3d58b5bd45643bbefeaeaafd97 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-a, io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.license=GPLv2, ceph=True, org.label-schema.vendor=CentOS) 2026-03-09T17:41:09.601 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 podman[52712]: 2026-03-09 17:41:09.555299341 +0000 UTC m=+0.291120108 container remove 3d6814385085dd8644a6824b89f9dbdb64d0ee3d58b5bd45643bbefeaeaafd97 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-a, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223, CEPH_REF=squid) 2026-03-09T17:41:09.601 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 bash[52712]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-a 2026-03-09T17:41:09.775 INFO:teuthology.orchestra.run.vm01.stdout:Setting public_network to 192.168.123.0/24 in mon config section 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.a.service: Deactivated successfully. 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 systemd[1]: Stopped Ceph mon.a for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 systemd[1]: Starting Ceph mon.a for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 podman[52779]: 2026-03-09 17:41:09.7249619 +0000 UTC m=+0.022442169 container create 98555f31d8dac30bb086772aab311a7e00bf9d59c18ac380dd9f5afca74e4afc (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-a, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, CEPH_REF=squid, ceph=True, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 podman[52779]: 2026-03-09 17:41:09.758984674 +0000 UTC m=+0.056464943 container init 98555f31d8dac30bb086772aab311a7e00bf9d59c18ac380dd9f5afca74e4afc (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-a, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, ceph=True, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 podman[52779]: 2026-03-09 17:41:09.762586406 +0000 UTC m=+0.060066665 container start 98555f31d8dac30bb086772aab311a7e00bf9d59c18ac380dd9f5afca74e4afc (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-a, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, org.label-schema.license=GPLv2, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 bash[52779]: 98555f31d8dac30bb086772aab311a7e00bf9d59c18ac380dd9f5afca74e4afc 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 podman[52779]: 2026-03-09 17:41:09.714549329 +0000 UTC m=+0.012029607 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 systemd[1]: Started Ceph mon.a for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: set uid:gid to 167:167 (ceph:ceph) 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 2 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: pidfile_write: ignore empty --pid-file 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: load: jerasure load: lrc 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: RocksDB version: 7.9.2 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Git sha 0 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Compile date 2026-02-25 18:11:04 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: DB SUMMARY 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: DB Session ID: MTHQH7JDHCU4RR0RWVIH 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: CURRENT file: CURRENT 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: IDENTITY file: IDENTITY 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 75519 ; 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.error_if_exists: 0 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.create_if_missing: 0 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.paranoid_checks: 1 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.env: 0x55856dd01dc0 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.fs: PosixFileSystem 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.info_log: 0x55856fd48700 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_file_opening_threads: 16 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.statistics: (nil) 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.use_fsync: 0 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_log_file_size: 0 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.keep_log_file_num: 1000 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.recycle_log_file_num: 0 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.allow_fallocate: 1 2026-03-09T17:41:09.853 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.allow_mmap_reads: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.allow_mmap_writes: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.use_direct_reads: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.create_missing_column_families: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.db_log_dir: 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.wal_dir: 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.advise_random_on_open: 1 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.db_write_buffer_size: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.write_buffer_manager: 0x55856fd4d900 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.rate_limiter: (nil) 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.wal_recovery_mode: 2 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.enable_thread_tracking: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.enable_pipelined_write: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.unordered_write: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.row_cache: None 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.wal_filter: None 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.allow_ingest_behind: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.two_write_queues: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.manual_wal_flush: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.wal_compression: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.atomic_flush: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.log_readahead_size: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.best_efforts_recovery: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.allow_data_in_errors: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.db_host_id: __hostname__ 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_background_jobs: 2 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_background_compactions: -1 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_subcompactions: 1 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_total_wal_size: 0 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-09T17:41:09.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_open_files: -1 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.bytes_per_sync: 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compaction_readahead_size: 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_background_flushes: -1 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Compression algorithms supported: 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: kZSTD supported: 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: kXpressCompression supported: 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: kBZip2Compression supported: 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: kLZ4Compression supported: 1 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: kZlibCompression supported: 1 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: kLZ4HCCompression supported: 1 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: kSnappyCompression supported: 1 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.merge_operator: 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compaction_filter: None 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compaction_filter_factory: None 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.sst_partitioner_factory: None 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55856fd48640) 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: cache_index_and_filter_blocks: 1 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: pin_top_level_index_and_filter: 1 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: index_type: 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: data_block_index_type: 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: index_shortening: 1 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: checksum: 4 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: no_block_cache: 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: block_cache: 0x55856fd6d350 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: block_cache_name: BinnedLRUCache 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: block_cache_options: 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: capacity : 536870912 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: num_shard_bits : 4 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: strict_capacity_limit : 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: high_pri_pool_ratio: 0.000 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: block_cache_compressed: (nil) 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: persistent_cache: (nil) 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: block_size: 4096 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: block_size_deviation: 10 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: block_restart_interval: 16 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: index_block_restart_interval: 1 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: metadata_block_size: 4096 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: partition_filters: 0 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: use_delta_encoding: 1 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: filter_policy: bloomfilter 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: whole_key_filtering: 1 2026-03-09T17:41:09.855 INFO:journalctl@ceph.mon.a.vm01.stdout: verify_compression: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout: read_amp_bytes_per_bit: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout: format_version: 5 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout: enable_index_compression: 1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout: block_align: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout: max_auto_readahead_size: 262144 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout: prepopulate_block_cache: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout: initial_auto_readahead_size: 8192 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout: num_file_reads_for_auto_readahead: 2 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.write_buffer_size: 33554432 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_write_buffer_number: 2 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compression: NoCompression 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.bottommost_compression: Disabled 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.prefix_extractor: nullptr 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.num_levels: 7 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compression_opts.level: 32767 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compression_opts.strategy: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compression_opts.enabled: false 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.target_file_size_base: 67108864 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.arena_block_size: 1048576 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.disable_auto_compactions: 0 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-09T17:41:09.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.inplace_update_support: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.bloom_locality: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.max_successive_merges: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.paranoid_file_checks: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.force_consistency_checks: 1 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.report_bg_io_stats: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.ttl: 2592000 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.enable_blob_files: false 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.min_blob_size: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.blob_file_size: 268435456 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.blob_file_starting_level: 0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 16e27686-986c-49ec-8a05-2e3de77d683e 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773078069803226, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773078069806332, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 72589, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 225, "table_properties": {"data_size": 70864, "index_size": 178, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9693, "raw_average_key_size": 49, "raw_value_size": 65370, "raw_average_value_size": 333, "num_data_blocks": 8, "num_entries": 196, "num_filter_entries": 196, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773078069, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "16e27686-986c-49ec-8a05-2e3de77d683e", "db_session_id": "MTHQH7JDHCU4RR0RWVIH", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773078069806416, "job": 1, "event": "recovery_finished"} 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55856fd6ee00 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: DB pointer 0x55856fe84000 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: ** DB Stats ** 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: ** Compaction Stats [default] ** 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: L0 2/0 72.74 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 23.6 0.00 0.00 1 0.003 0 0 0.0 0.0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: Sum 2/0 72.74 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 23.6 0.00 0.00 1 0.003 0 0 0.0 0.0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 23.6 0.00 0.00 1 0.003 0 0 0.0 0.0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: ** Compaction Stats [default] ** 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 23.6 0.00 0.00 1 0.003 0 0 0.0 0.0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T17:41:09.857 INFO:journalctl@ceph.mon.a.vm01.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: Cumulative compaction: 0.00 GB write, 6.61 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: Interval compaction: 0.00 GB write, 6.61 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: Block cache BinnedLRUCache@0x55856fd6d350#2 capacity: 512.00 MB usage: 1.08 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1e-05 secs_since: 0 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: Block cache entry stats(count,size,portion): FilterBlock(2,0.70 KB,0.00013411%) IndexBlock(2,0.38 KB,7.15256e-05%) Misc(1,0.00 KB,0%) 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: starting mon.a rank 0 at public addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] at bind addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: mon.a@-1(???) e1 preinit fsid 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: mon.a@-1(???).mds e1 new map 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: mon.a@-1(???).mds e1 print_map 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: e1 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: btime 2026-03-09T17:41:08:650714+0000 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: legacy client fscid: -1 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout: No filesystems configured 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: mon.a@-1(???).osd e1 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T17:41:09.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:09 vm01 ceph-mon[52793]: mon.a@-1(???).paxosservice(auth 1..2) refresh upgraded, format 0 -> 3 2026-03-09T17:41:09.968 INFO:teuthology.orchestra.run.vm01.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-09T17:41:09.968 INFO:teuthology.orchestra.run.vm01.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-09T17:41:09.968 INFO:teuthology.orchestra.run.vm01.stdout:Creating mgr... 2026-03-09T17:41:09.969 INFO:teuthology.orchestra.run.vm01.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-09T17:41:09.969 INFO:teuthology.orchestra.run.vm01.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-09T17:41:10.130 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-01455850-1bdf-11f1-910a-9936d43313cc@mgr.x 2026-03-09T17:41:10.130 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to reset failed state of unit ceph-01455850-1bdf-11f1-910a-9936d43313cc@mgr.x.service: Unit ceph-01455850-1bdf-11f1-910a-9936d43313cc@mgr.x.service not loaded. 2026-03-09T17:41:10.272 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-01455850-1bdf-11f1-910a-9936d43313cc.target.wants/ceph-01455850-1bdf-11f1-910a-9936d43313cc@mgr.x.service → /etc/systemd/system/ceph-01455850-1bdf-11f1-910a-9936d43313cc@.service. 2026-03-09T17:41:10.451 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-09T17:41:10.451 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to enable service . firewalld.service is not available 2026-03-09T17:41:10.451 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-09T17:41:10.451 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-09T17:41:10.451 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr to start... 2026-03-09T17:41:10.451 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr... 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "01455850-1bdf-11f1-910a-9936d43313cc", 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 0, 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T17:41:10.695 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T17:41:08:650714+0000", 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T17:41:08.652472+0000", 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-09T17:41:10.696 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (1/15)... 2026-03-09T17:41:10.857 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:10 vm01 ceph-mon[52793]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-09T17:41:10.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:10 vm01 ceph-mon[52793]: monmap epoch 1 2026-03-09T17:41:10.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:10 vm01 ceph-mon[52793]: fsid 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:10.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:10 vm01 ceph-mon[52793]: last_changed 2026-03-09T17:41:07.435708+0000 2026-03-09T17:41:10.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:10 vm01 ceph-mon[52793]: created 2026-03-09T17:41:07.435708+0000 2026-03-09T17:41:10.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:10 vm01 ceph-mon[52793]: min_mon_release 19 (squid) 2026-03-09T17:41:10.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:10 vm01 ceph-mon[52793]: election_strategy: 1 2026-03-09T17:41:10.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:10 vm01 ceph-mon[52793]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T17:41:10.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:10 vm01 ceph-mon[52793]: fsmap 2026-03-09T17:41:10.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:10 vm01 ceph-mon[52793]: osdmap e1: 0 total, 0 up, 0 in 2026-03-09T17:41:10.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:10 vm01 ceph-mon[52793]: mgrmap e1: no daemons active 2026-03-09T17:41:10.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:10 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/4153742840' entity='client.admin' 2026-03-09T17:41:10.858 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:10 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2082410158' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "01455850-1bdf-11f1-910a-9936d43313cc", 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T17:41:12.983 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T17:41:08:650714+0000", 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T17:41:08.652472+0000", 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-09T17:41:12.984 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (2/15)... 2026-03-09T17:41:13.123 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:13 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/930253623' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T17:41:14.130 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:14 vm01 ceph-mon[52793]: Activating manager daemon x 2026-03-09T17:41:14.131 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:14 vm01 ceph-mon[52793]: mgrmap e2: x(active, starting, since 0.00384905s) 2026-03-09T17:41:14.131 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:14 vm01 ceph-mon[52793]: from='mgr.14100 192.168.123.101:0/1899620653' entity='mgr.x' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T17:41:14.131 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:14 vm01 ceph-mon[52793]: from='mgr.14100 192.168.123.101:0/1899620653' entity='mgr.x' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T17:41:14.131 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:14 vm01 ceph-mon[52793]: from='mgr.14100 192.168.123.101:0/1899620653' entity='mgr.x' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T17:41:14.131 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:14 vm01 ceph-mon[52793]: from='mgr.14100 192.168.123.101:0/1899620653' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:41:14.131 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:14 vm01 ceph-mon[52793]: from='mgr.14100 192.168.123.101:0/1899620653' entity='mgr.x' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-09T17:41:14.131 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:14 vm01 ceph-mon[52793]: Manager daemon x is now available 2026-03-09T17:41:14.131 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:14 vm01 ceph-mon[52793]: from='mgr.14100 192.168.123.101:0/1899620653' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:41:14.131 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:14 vm01 ceph-mon[52793]: from='mgr.14100 192.168.123.101:0/1899620653' entity='mgr.x' 2026-03-09T17:41:14.131 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:14 vm01 ceph-mon[52793]: from='mgr.14100 192.168.123.101:0/1899620653' entity='mgr.x' 2026-03-09T17:41:14.131 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:14 vm01 ceph-mon[52793]: from='mgr.14100 192.168.123.101:0/1899620653' entity='mgr.x' 2026-03-09T17:41:14.131 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:14 vm01 ceph-mon[52793]: from='mgr.14100 192.168.123.101:0/1899620653' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"}]: dispatch 2026-03-09T17:41:15.292 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T17:41:15.292 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-09T17:41:15.292 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "01455850-1bdf-11f1-910a-9936d43313cc", 2026-03-09T17:41:15.292 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T17:41:08:650714+0000", 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T17:41:08.652472+0000", 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-09T17:41:15.293 INFO:teuthology.orchestra.run.vm01.stdout:mgr is available 2026-03-09T17:41:15.563 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T17:41:15.563 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [global] 2026-03-09T17:41:15.563 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout fsid = 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:15.564 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-09T17:41:15.564 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-09T17:41:15.564 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-09T17:41:15.564 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-09T17:41:15.564 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-09T17:41:15.564 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-09T17:41:15.564 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T17:41:15.564 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-09T17:41:15.564 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-09T17:41:15.564 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-09T17:41:15.564 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [osd] 2026-03-09T17:41:15.564 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-09T17:41:15.564 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-09T17:41:15.564 INFO:teuthology.orchestra.run.vm01.stdout:Enabling cephadm module... 2026-03-09T17:41:15.902 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:15 vm01 ceph-mon[52793]: mgrmap e3: x(active, since 1.00927s) 2026-03-09T17:41:15.902 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:15 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2639600417' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T17:41:15.902 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:15 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3335081534' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-09T17:41:16.910 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:16 vm01 ceph-mon[52793]: mgrmap e4: x(active, since 2s) 2026-03-09T17:41:16.910 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:16 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3839830164' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-09T17:41:17.132 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-09T17:41:17.132 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 5, 2026-03-09T17:41:17.132 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T17:41:17.132 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "active_name": "x", 2026-03-09T17:41:17.132 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-09T17:41:17.132 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-09T17:41:17.132 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for the mgr to restart... 2026-03-09T17:41:17.132 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr epoch 5... 2026-03-09T17:41:17.920 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:17 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3839830164' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-09T17:41:17.920 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:17 vm01 ceph-mon[52793]: mgrmap e5: x(active, since 3s) 2026-03-09T17:41:17.920 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:17 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/568257366' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T17:41:18.933 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:18 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:18.927+0000 7fa3878c4140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T17:41:19.184 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:18 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:18.969+0000 7fa3878c4140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T17:41:19.184 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:19 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:19.008+0000 7fa3878c4140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T17:41:19.435 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:19 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:19.184+0000 7fa3878c4140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T17:41:19.436 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:19 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:19.237+0000 7fa3878c4140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T17:41:19.687 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:19 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:19.472+0000 7fa3878c4140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T17:41:19.940 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:19 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:19.793+0000 7fa3878c4140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T17:41:19.940 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:19 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:19.837+0000 7fa3878c4140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T17:41:19.940 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:19 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:19.886+0000 7fa3878c4140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T17:41:20.190 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:19 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:19.973+0000 7fa3878c4140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T17:41:20.190 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:20 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:20.015+0000 7fa3878c4140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T17:41:20.190 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:20 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:20.107+0000 7fa3878c4140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T17:41:20.440 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:20 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:20.242+0000 7fa3878c4140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T17:41:20.440 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:20 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:20.400+0000 7fa3878c4140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T17:41:20.691 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:20 vm01 ceph-mon[52793]: Active manager daemon x restarted 2026-03-09T17:41:20.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:20 vm01 ceph-mon[52793]: Activating manager daemon x 2026-03-09T17:41:20.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:20 vm01 ceph-mon[52793]: osdmap e2: 0 total, 0 up, 0 in 2026-03-09T17:41:20.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:20 vm01 ceph-mon[52793]: mgrmap e6: x(active, starting, since 0.00908522s) 2026-03-09T17:41:20.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:20 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:41:20.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:20 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-09T17:41:20.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:20 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T17:41:20.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:20 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T17:41:20.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:20 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T17:41:20.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:20 vm01 ceph-mon[52793]: Manager daemon x is now available 2026-03-09T17:41:20.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:20 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' 2026-03-09T17:41:20.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:20 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' 2026-03-09T17:41:20.692 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:20 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:20.444+0000 7fa3878c4140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T17:41:21.543 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-09T17:41:21.543 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 7, 2026-03-09T17:41:21.543 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-09T17:41:21.543 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-09T17:41:21.543 INFO:teuthology.orchestra.run.vm01.stdout:mgr epoch 5 is available 2026-03-09T17:41:21.543 INFO:teuthology.orchestra.run.vm01.stdout:Setting orchestrator backend to cephadm... 2026-03-09T17:41:21.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:21 vm01 ceph-mon[52793]: Found migration_current of "None". Setting to last migration. 2026-03-09T17:41:21.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:21 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:21.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:21 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:41:21.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:21 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:21.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:21 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"}]: dispatch 2026-03-09T17:41:21.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:21 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' 2026-03-09T17:41:21.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:21 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' 2026-03-09T17:41:21.699 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:21 vm01 ceph-mon[52793]: mgrmap e7: x(active, since 1.01192s) 2026-03-09T17:41:22.315 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-09T17:41:22.315 INFO:teuthology.orchestra.run.vm01.stdout:Generating ssh key... 2026-03-09T17:41:22.710 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: Generating public/private ed25519 key pair. 2026-03-09T17:41:22.710 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: Your identification has been saved in /tmp/tmp3afxapcu/key 2026-03-09T17:41:22.710 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: Your public key has been saved in /tmp/tmp3afxapcu/key.pub 2026-03-09T17:41:22.710 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: The key fingerprint is: 2026-03-09T17:41:22.710 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: SHA256:eCFbtVEkOszPhfb5G4cYGHUQ0Df7h+Q+LbjYt7BRzA0 ceph-01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:22.710 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: The key's randomart image is: 2026-03-09T17:41:22.710 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: +--[ED25519 256]--+ 2026-03-09T17:41:22.710 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: | ==*o. | 2026-03-09T17:41:22.710 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: | o o *..o | 2026-03-09T17:41:22.710 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: | . B = ..Eo | 2026-03-09T17:41:22.711 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: | = * = +oo | 2026-03-09T17:41:22.711 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: | o S + +o+o.| 2026-03-09T17:41:22.711 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: | . =o.o| 2026-03-09T17:41:22.711 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: | +o+.o| 2026-03-09T17:41:22.711 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: | o.+=+.| 2026-03-09T17:41:22.711 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: | . +oo+ | 2026-03-09T17:41:22.711 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:22 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: +----[SHA256]-----+ 2026-03-09T17:41:22.925 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIPqzlMs1TV84ORYo6Gz8vwMXK1qob/20l3eSmrpQSzP ceph-01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:22.925 INFO:teuthology.orchestra.run.vm01.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-09T17:41:22.925 INFO:teuthology.orchestra.run.vm01.stdout:Adding key to root@localhost authorized_keys... 2026-03-09T17:41:22.926 INFO:teuthology.orchestra.run.vm01.stdout:Adding host vm01... 2026-03-09T17:41:22.964 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:22 vm01 ceph-mon[52793]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T17:41:22.965 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:22 vm01 ceph-mon[52793]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T17:41:22.965 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:22 vm01 ceph-mon[52793]: [09/Mar/2026:17:41:21] ENGINE Bus STARTING 2026-03-09T17:41:22.965 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:22 vm01 ceph-mon[52793]: [09/Mar/2026:17:41:21] ENGINE Serving on https://192.168.123.101:7150 2026-03-09T17:41:22.965 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:22 vm01 ceph-mon[52793]: [09/Mar/2026:17:41:21] ENGINE Client ('192.168.123.101', 49106) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T17:41:22.965 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:22 vm01 ceph-mon[52793]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:41:22.965 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:22 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' 2026-03-09T17:41:22.965 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:22 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:22.965 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:22 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:22.965 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:22 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' 2026-03-09T17:41:22.965 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:22 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' 2026-03-09T17:41:23.974 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:23 vm01 ceph-mon[52793]: [09/Mar/2026:17:41:21] ENGINE Serving on http://192.168.123.101:8765 2026-03-09T17:41:23.974 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:23 vm01 ceph-mon[52793]: [09/Mar/2026:17:41:21] ENGINE Bus STARTED 2026-03-09T17:41:23.974 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:23 vm01 ceph-mon[52793]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:41:23.974 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:23 vm01 ceph-mon[52793]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:41:23.974 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:23 vm01 ceph-mon[52793]: Generating ssh key... 2026-03-09T17:41:23.974 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:23 vm01 ceph-mon[52793]: mgrmap e8: x(active, since 2s) 2026-03-09T17:41:24.892 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Added host 'vm01' with addr '192.168.123.101' 2026-03-09T17:41:24.892 INFO:teuthology.orchestra.run.vm01.stdout:Deploying unmanaged mon service... 2026-03-09T17:41:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:24 vm01 ceph-mon[52793]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:41:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:24 vm01 ceph-mon[52793]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm01", "addr": "192.168.123.101", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:41:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:24 vm01 ceph-mon[52793]: Deploying cephadm binary to vm01 2026-03-09T17:41:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:24 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' 2026-03-09T17:41:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:24 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:25.317 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-09T17:41:25.317 INFO:teuthology.orchestra.run.vm01.stdout:Deploying unmanaged mgr service... 2026-03-09T17:41:25.627 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-09T17:41:26.684 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:26 vm01 ceph-mon[52793]: Added host vm01 2026-03-09T17:41:26.685 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:26 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' 2026-03-09T17:41:26.685 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:26 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' 2026-03-09T17:41:26.718 INFO:teuthology.orchestra.run.vm01.stdout:Enabling the dashboard module... 2026-03-09T17:41:27.398 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:27 vm01 ceph-mon[52793]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:41:27.399 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:27 vm01 ceph-mon[52793]: Saving service mon spec with placement count:5 2026-03-09T17:41:27.399 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:27 vm01 ceph-mon[52793]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:41:27.399 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:27 vm01 ceph-mon[52793]: Saving service mgr spec with placement count:2 2026-03-09T17:41:27.399 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:27 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1668540644' entity='client.admin' 2026-03-09T17:41:27.399 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:27 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' 2026-03-09T17:41:27.399 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:27 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2142000167' entity='client.admin' 2026-03-09T17:41:27.399 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:27 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' 2026-03-09T17:41:27.399 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:27 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/155945003' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-09T17:41:27.399 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:27 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' 2026-03-09T17:41:28.232 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:28 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: ignoring --setuser ceph since I am not root 2026-03-09T17:41:28.232 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:28 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: ignoring --setgroup ceph since I am not root 2026-03-09T17:41:28.232 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:28 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:28.210+0000 7f761bbc5140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T17:41:28.430 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-09T17:41:28.430 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 9, 2026-03-09T17:41:28.430 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T17:41:28.430 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "active_name": "x", 2026-03-09T17:41:28.430 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-09T17:41:28.430 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-09T17:41:28.430 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for the mgr to restart... 2026-03-09T17:41:28.430 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr epoch 9... 2026-03-09T17:41:28.509 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:28 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:28.259+0000 7f761bbc5140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T17:41:28.978 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:28 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:28.714+0000 7f761bbc5140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T17:41:28.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:28 vm01 ceph-mon[52793]: from='mgr.14118 192.168.123.101:0/3173588674' entity='mgr.x' 2026-03-09T17:41:28.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:28 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/155945003' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-09T17:41:28.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:28 vm01 ceph-mon[52793]: mgrmap e9: x(active, since 7s) 2026-03-09T17:41:28.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:28 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/853353804' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T17:41:29.478 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:29 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:29.051+0000 7f761bbc5140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T17:41:29.478 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:29 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T17:41:29.478 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:29 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T17:41:29.478 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:29 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: from numpy import show_config as show_numpy_config 2026-03-09T17:41:29.478 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:29 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:29.141+0000 7f761bbc5140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T17:41:29.478 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:29 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:29.178+0000 7f761bbc5140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T17:41:29.478 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:29 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:29.252+0000 7f761bbc5140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T17:41:30.038 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:29 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:29.764+0000 7f761bbc5140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T17:41:30.038 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:29 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:29.877+0000 7f761bbc5140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T17:41:30.038 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:29 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:29.920+0000 7f761bbc5140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T17:41:30.038 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:29 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:29.955+0000 7f761bbc5140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T17:41:30.038 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:29 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:29.998+0000 7f761bbc5140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T17:41:30.478 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:30 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:30.036+0000 7f761bbc5140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T17:41:30.478 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:30 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:30.211+0000 7f761bbc5140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T17:41:30.478 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:30 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:30.264+0000 7f761bbc5140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T17:41:30.779 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:30 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:30.490+0000 7f761bbc5140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T17:41:31.062 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:30 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:30.778+0000 7f761bbc5140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T17:41:31.062 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:30 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:30.816+0000 7f761bbc5140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T17:41:31.063 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:30 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:30.859+0000 7f761bbc5140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T17:41:31.063 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:30 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:30.939+0000 7f761bbc5140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T17:41:31.063 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:30 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:30.979+0000 7f761bbc5140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T17:41:31.322 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:31 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:31.061+0000 7f761bbc5140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T17:41:31.323 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:31 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:31.180+0000 7f761bbc5140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:31 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:31.321+0000 7f761bbc5140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:31 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:31.360+0000 7f761bbc5140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:31 vm01 ceph-mon[52793]: Active manager daemon x restarted 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:31 vm01 ceph-mon[52793]: Activating manager daemon x 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:31 vm01 ceph-mon[52793]: osdmap e3: 0 total, 0 up, 0 in 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:31 vm01 ceph-mon[52793]: mgrmap e10: x(active, starting, since 0.00991079s) 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:31 vm01 ceph-mon[52793]: Manager daemon x is now available 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:41:31.729 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"}]: dispatch 2026-03-09T17:41:32.434 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-09T17:41:32.434 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 11, 2026-03-09T17:41:32.434 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-09T17:41:32.434 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-09T17:41:32.434 INFO:teuthology.orchestra.run.vm01.stdout:mgr epoch 9 is available 2026-03-09T17:41:32.434 INFO:teuthology.orchestra.run.vm01.stdout:Generating a dashboard self-signed certificate... 2026-03-09T17:41:32.785 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-09T17:41:32.785 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial admin user... 2026-03-09T17:41:33.243 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:33 vm01 ceph-mon[52793]: [09/Mar/2026:17:41:32] ENGINE Bus STARTING 2026-03-09T17:41:33.243 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:33.243 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:33 vm01 ceph-mon[52793]: [09/Mar/2026:17:41:32] ENGINE Serving on https://192.168.123.101:7150 2026-03-09T17:41:33.243 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:33 vm01 ceph-mon[52793]: [09/Mar/2026:17:41:32] ENGINE Client ('192.168.123.101', 55290) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T17:41:33.243 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:33 vm01 ceph-mon[52793]: [09/Mar/2026:17:41:32] ENGINE Serving on http://192.168.123.101:8765 2026-03-09T17:41:33.243 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:33 vm01 ceph-mon[52793]: [09/Mar/2026:17:41:32] ENGINE Bus STARTED 2026-03-09T17:41:33.243 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:33 vm01 ceph-mon[52793]: mgrmap e11: x(active, since 1.01166s) 2026-03-09T17:41:33.243 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:33 vm01 ceph-mon[52793]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T17:41:33.243 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:33 vm01 ceph-mon[52793]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T17:41:33.243 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:33 vm01 ceph-mon[52793]: from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:41:33.243 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:33.243 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:33.243 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:33.243 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:33 vm01 ceph-mon[52793]: from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:41:33.264 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$xU6Hl.GCSvT3VtG.RaCbd.IDWoSHfffvVzAS2TMD9nx8uRlMuTO7m", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773078093, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-09T17:41:33.265 INFO:teuthology.orchestra.run.vm01.stdout:Fetching dashboard port number... 2026-03-09T17:41:33.510 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 8443 2026-03-09T17:41:33.510 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-09T17:41:33.510 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-09T17:41:33.512 INFO:teuthology.orchestra.run.vm01.stdout:Ceph Dashboard is now available at: 2026-03-09T17:41:33.512 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:41:33.512 INFO:teuthology.orchestra.run.vm01.stdout: URL: https://vm01.local:8443/ 2026-03-09T17:41:33.512 INFO:teuthology.orchestra.run.vm01.stdout: User: admin 2026-03-09T17:41:33.512 INFO:teuthology.orchestra.run.vm01.stdout: Password: zc7795d1v9 2026-03-09T17:41:33.512 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:41:33.512 INFO:teuthology.orchestra.run.vm01.stdout:Saving cluster configuration to /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config directory 2026-03-09T17:41:33.828 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-09T17:41:33.828 INFO:teuthology.orchestra.run.vm01.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-09T17:41:33.828 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:41:33.828 INFO:teuthology.orchestra.run.vm01.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-09T17:41:33.828 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:41:33.828 INFO:teuthology.orchestra.run.vm01.stdout:Or, if you are only running a single cluster on this host: 2026-03-09T17:41:33.828 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:41:33.828 INFO:teuthology.orchestra.run.vm01.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-09T17:41:33.828 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:41:33.828 INFO:teuthology.orchestra.run.vm01.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-09T17:41:33.828 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:41:33.828 INFO:teuthology.orchestra.run.vm01.stdout: ceph telemetry on 2026-03-09T17:41:33.828 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:41:33.828 INFO:teuthology.orchestra.run.vm01.stdout:For more information see: 2026-03-09T17:41:33.829 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:41:33.829 INFO:teuthology.orchestra.run.vm01.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-09T17:41:33.829 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:41:33.829 INFO:teuthology.orchestra.run.vm01.stdout:Bootstrap complete. 2026-03-09T17:41:33.858 INFO:tasks.cephadm:Fetching config... 2026-03-09T17:41:33.858 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:41:33.858 DEBUG:teuthology.orchestra.run.vm01:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-09T17:41:33.888 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-09T17:41:33.888 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:41:33.888 DEBUG:teuthology.orchestra.run.vm01:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-09T17:41:33.950 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-09T17:41:33.950 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:41:33.950 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/keyring of=/dev/stdout 2026-03-09T17:41:34.019 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-09T17:41:34.020 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:41:34.020 DEBUG:teuthology.orchestra.run.vm01:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-09T17:41:34.075 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-09T17:41:34.075 DEBUG:teuthology.orchestra.run.vm01:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIPqzlMs1TV84ORYo6Gz8vwMXK1qob/20l3eSmrpQSzP ceph-01455850-1bdf-11f1-910a-9936d43313cc' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T17:41:34.169 INFO:teuthology.orchestra.run.vm01.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIPqzlMs1TV84ORYo6Gz8vwMXK1qob/20l3eSmrpQSzP ceph-01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:34.183 DEBUG:teuthology.orchestra.run.vm04:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIPqzlMs1TV84ORYo6Gz8vwMXK1qob/20l3eSmrpQSzP ceph-01455850-1bdf-11f1-910a-9936d43313cc' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T17:41:34.221 INFO:teuthology.orchestra.run.vm04.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIPqzlMs1TV84ORYo6Gz8vwMXK1qob/20l3eSmrpQSzP ceph-01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:34.233 DEBUG:teuthology.orchestra.run.vm05:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIPqzlMs1TV84ORYo6Gz8vwMXK1qob/20l3eSmrpQSzP ceph-01455850-1bdf-11f1-910a-9936d43313cc' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T17:41:34.271 INFO:teuthology.orchestra.run.vm05.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIIPqzlMs1TV84ORYo6Gz8vwMXK1qob/20l3eSmrpQSzP ceph-01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:34.282 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-09T17:41:34.463 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:41:34.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:34 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:34.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:34 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/986235572' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-09T17:41:34.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:34 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2600182109' entity='client.admin' 2026-03-09T17:41:34.932 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-09T17:41:34.932 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-09T17:41:35.151 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:41:35.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:35 vm01 ceph-mon[52793]: mgrmap e12: x(active, since 2s) 2026-03-09T17:41:35.425 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:35 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2133704698' entity='client.admin' 2026-03-09T17:41:35.464 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm04 2026-03-09T17:41:35.464 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:41:35.464 DEBUG:teuthology.orchestra.run.vm04:> dd of=/etc/ceph/ceph.conf 2026-03-09T17:41:35.481 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:41:35.481 DEBUG:teuthology.orchestra.run.vm04:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:41:35.538 INFO:tasks.cephadm:Adding host vm04 to orchestrator... 2026-03-09T17:41:35.538 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch host add vm04 2026-03-09T17:41:35.722 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:41:36.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:36 vm01 ceph-mon[52793]: from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:41:36.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:36.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:36 vm01 ceph-mon[52793]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm04", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:41:36.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:36.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:36.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:41:36.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:36.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:36.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:36 vm01 ceph-mon[52793]: Updating vm01:/etc/ceph/ceph.conf 2026-03-09T17:41:37.607 INFO:teuthology.orchestra.run.vm01.stdout:Added host 'vm04' with addr '192.168.123.104' 2026-03-09T17:41:37.681 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch host ls --format=json 2026-03-09T17:41:37.866 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:41:37.889 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:37 vm01 ceph-mon[52793]: Updating vm01:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.conf 2026-03-09T17:41:37.889 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:37 vm01 ceph-mon[52793]: Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:41:37.889 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:37 vm01 ceph-mon[52793]: Deploying cephadm binary to vm04 2026-03-09T17:41:37.889 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:37 vm01 ceph-mon[52793]: Updating vm01:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.client.admin.keyring 2026-03-09T17:41:37.889 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:37.889 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:37.889 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:37.889 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:37.889 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:37.889 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:37.889 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:37.889 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:37.889 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:38.110 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:41:38.110 INFO:teuthology.orchestra.run.vm01.stdout:[{"addr": "192.168.123.101", "hostname": "vm01", "labels": [], "status": ""}, {"addr": "192.168.123.104", "hostname": "vm04", "labels": [], "status": ""}] 2026-03-09T17:41:38.162 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm05 2026-03-09T17:41:38.162 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-09T17:41:38.162 DEBUG:teuthology.orchestra.run.vm05:> dd of=/etc/ceph/ceph.conf 2026-03-09T17:41:38.178 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-09T17:41:38.178 DEBUG:teuthology.orchestra.run.vm05:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:41:38.236 INFO:tasks.cephadm:Adding host vm05 to orchestrator... 2026-03-09T17:41:38.237 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch host add vm05 2026-03-09T17:41:38.409 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:41:39.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:38 vm01 ceph-mon[52793]: Added host vm04 2026-03-09T17:41:39.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:38 vm01 ceph-mon[52793]: mgrmap e13: x(active, since 6s) 2026-03-09T17:41:39.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:39.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:40.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:39 vm01 ceph-mon[52793]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:41:40.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:39 vm01 ceph-mon[52793]: from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm05", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:41:40.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:39 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:40.260 INFO:teuthology.orchestra.run.vm01.stdout:Added host 'vm05' with addr '192.168.123.105' 2026-03-09T17:41:40.317 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch host ls --format=json 2026-03-09T17:41:40.498 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:41:40.738 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:41:40.738 INFO:teuthology.orchestra.run.vm01.stdout:[{"addr": "192.168.123.101", "hostname": "vm01", "labels": [], "status": ""}, {"addr": "192.168.123.104", "hostname": "vm04", "labels": [], "status": ""}, {"addr": "192.168.123.105", "hostname": "vm05", "labels": [], "status": ""}] 2026-03-09T17:41:40.789 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-09T17:41:40.790 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd crush tunables default 2026-03-09T17:41:40.973 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:41:41.039 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:40 vm01 ceph-mon[52793]: Deploying cephadm binary to vm05 2026-03-09T17:41:41.039 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:41.039 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:41.039 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:41.039 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:41.039 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:41.039 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:41:41.039 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:41.039 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:41.039 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:41.039 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:41.039 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:42.090 INFO:teuthology.orchestra.run.vm01.stderr:adjusted tunables profile to default 2026-03-09T17:41:42.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:41 vm01 ceph-mon[52793]: Added host vm05 2026-03-09T17:41:42.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:41 vm01 ceph-mon[52793]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T17:41:42.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:41 vm01 ceph-mon[52793]: Updating vm04:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.conf 2026-03-09T17:41:42.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:41 vm01 ceph-mon[52793]: Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:41:42.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:41 vm01 ceph-mon[52793]: Updating vm04:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.client.admin.keyring 2026-03-09T17:41:42.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:41 vm01 ceph-mon[52793]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:41:42.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:41 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:42.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:41 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:42.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:41 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3215700579' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-09T17:41:42.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:41 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:42.145 INFO:tasks.cephadm:Adding mon.a on vm01 2026-03-09T17:41:42.145 INFO:tasks.cephadm:Adding mon.b on vm04 2026-03-09T17:41:42.145 INFO:tasks.cephadm:Adding mon.c on vm05 2026-03-09T17:41:42.145 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch apply mon '3;vm01:192.168.123.101=a;vm04:192.168.123.104=b;vm05:192.168.123.105=c' 2026-03-09T17:41:42.339 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T17:41:42.386 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T17:41:42.646 INFO:teuthology.orchestra.run.vm05.stdout:Scheduled mon update... 2026-03-09T17:41:42.748 DEBUG:teuthology.orchestra.run.vm04:mon.b> sudo journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.b.service 2026-03-09T17:41:42.750 DEBUG:teuthology.orchestra.run.vm05:mon.c> sudo journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.c.service 2026-03-09T17:41:42.752 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-09T17:41:42.752 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph mon dump -f json 2026-03-09T17:41:42.999 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T17:41:43.053 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T17:41:43.384 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:41:43.384 INFO:teuthology.orchestra.run.vm05.stdout:{"epoch":1,"fsid":"01455850-1bdf-11f1-910a-9936d43313cc","modified":"2026-03-09T17:41:07.435708Z","created":"2026-03-09T17:41:07.435708Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:41:43.385 INFO:teuthology.orchestra.run.vm05.stderr:dumped monmap epoch 1 2026-03-09T17:41:43.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:43 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3215700579' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-09T17:41:43.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:43 vm01 ceph-mon[52793]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T17:41:43.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:43 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:43.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:43 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:44.470 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-09T17:41:44.470 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph mon dump -f json 2026-03-09T17:41:44.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:44 vm01 ceph-mon[52793]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm01:192.168.123.101=a;vm04:192.168.123.104=b;vm05:192.168.123.105=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:41:44.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:44 vm01 ceph-mon[52793]: Saving service mon spec with placement vm01:192.168.123.101=a;vm04:192.168.123.104=b;vm05:192.168.123.105=c;count:3 2026-03-09T17:41:44.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3273289525' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:41:44.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:44 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:44.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:44 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:44.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:44 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:44.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:44 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:44.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:44 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:41:44.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:44 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:44.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:44 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:44.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:44 vm01 ceph-mon[52793]: Updating vm05:/etc/ceph/ceph.conf 2026-03-09T17:41:44.673 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.conf 2026-03-09T17:41:44.979 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:41:44.979 INFO:teuthology.orchestra.run.vm05.stdout:{"epoch":1,"fsid":"01455850-1bdf-11f1-910a-9936d43313cc","modified":"2026-03-09T17:41:07.435708Z","created":"2026-03-09T17:41:07.435708Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:41:44.979 INFO:teuthology.orchestra.run.vm05.stderr:dumped monmap epoch 1 2026-03-09T17:41:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:45 vm01 ceph-mon[52793]: Updating vm05:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.conf 2026-03-09T17:41:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:45 vm01 ceph-mon[52793]: Updating vm05:/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:41:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:45 vm01 ceph-mon[52793]: Updating vm05:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.client.admin.keyring 2026-03-09T17:41:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:45 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:45 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:45 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:45 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:41:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:45 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:45 vm01 ceph-mon[52793]: Deploying daemon mon.c on vm05 2026-03-09T17:41:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:45 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/2248963214' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:41:46.059 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-09T17:41:46.059 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph mon dump -f json 2026-03-09T17:41:46.241 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.c/config 2026-03-09T17:41:47.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:47 vm04 ceph-mon[53158]: mon.b@-1(synchronizing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 3 2026-03-09T17:41:51.394 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:51 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: Deploying daemon mon.b on vm04 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: mon.a calling monitor election 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: mon.c calling monitor election 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: monmap epoch 2 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: fsid 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: last_changed 2026-03-09T17:41:46.050077+0000 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: created 2026-03-09T17:41:07.435708+0000 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: min_mon_release 19 (squid) 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: election_strategy: 1 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.c 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: fsmap 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: mgrmap e13: x(active, since 19s) 2026-03-09T17:41:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: overall HEALTH_OK 2026-03-09T17:41:51.479 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:51.479 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:51.479 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:51.479 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:52.478 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:52 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:52.048+0000 7f75e7f31640 -1 mgr.server handle_report got status from non-daemon mon.c 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: Deploying daemon mon.b on vm04 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: mon.a calling monitor election 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: mon.c calling monitor election 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: monmap epoch 2 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: fsid 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: last_changed 2026-03-09T17:41:46.050077+0000 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: created 2026-03-09T17:41:07.435708+0000 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: min_mon_release 19 (squid) 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: election_strategy: 1 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.c 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: fsmap 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: mgrmap e13: x(active, since 19s) 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: overall HEALTH_OK 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: mon.a calling monitor election 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: mon.c calling monitor election 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: mon.b calling monitor election 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: monmap epoch 3 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: fsid 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:56.793 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: last_changed 2026-03-09T17:41:51.495417+0000 2026-03-09T17:41:56.794 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: created 2026-03-09T17:41:07.435708+0000 2026-03-09T17:41:56.794 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: min_mon_release 19 (squid) 2026-03-09T17:41:56.794 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: election_strategy: 1 2026-03-09T17:41:56.794 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T17:41:56.794 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.c 2026-03-09T17:41:56.794 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T17:41:56.794 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: fsmap 2026-03-09T17:41:56.794 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T17:41:56.794 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: mgrmap e13: x(active, since 25s) 2026-03-09T17:41:56.794 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: overall HEALTH_OK 2026-03-09T17:41:56.794 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:56.794 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:56.794 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:56.794 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:56.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: mon.a calling monitor election 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: mon.c calling monitor election 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: mon.b calling monitor election 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: monmap epoch 3 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: fsid 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: last_changed 2026-03-09T17:41:51.495417+0000 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: created 2026-03-09T17:41:07.435708+0000 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: min_mon_release 19 (squid) 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: election_strategy: 1 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.c 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: fsmap 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: mgrmap e13: x(active, since 25s) 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: overall HEALTH_OK 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:56.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:56 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:56.924 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:41:56.924 INFO:teuthology.orchestra.run.vm05.stdout:{"epoch":3,"fsid":"01455850-1bdf-11f1-910a-9936d43313cc","modified":"2026-03-09T17:41:51.495417Z","created":"2026-03-09T17:41:07.435708Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:3300","nonce":0},{"type":"v1","addr":"192.168.123.105:6789","nonce":0}]},"addr":"192.168.123.105:6789/0","public_addr":"192.168.123.105:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":2,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:3300","nonce":0},{"type":"v1","addr":"192.168.123.104:6789","nonce":0}]},"addr":"192.168.123.104:6789/0","public_addr":"192.168.123.104:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1,2]} 2026-03-09T17:41:56.924 INFO:teuthology.orchestra.run.vm05.stderr:dumped monmap epoch 3 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: mon.a calling monitor election 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: mon.c calling monitor election 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: mon.b calling monitor election 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: monmap epoch 3 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: fsid 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: last_changed 2026-03-09T17:41:51.495417+0000 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: created 2026-03-09T17:41:07.435708+0000 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: min_mon_release 19 (squid) 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: election_strategy: 1 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-09T17:41:56.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.c 2026-03-09T17:41:56.980 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-09T17:41:56.980 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: fsmap 2026-03-09T17:41:56.980 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T17:41:56.980 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: mgrmap e13: x(active, since 25s) 2026-03-09T17:41:56.980 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: overall HEALTH_OK 2026-03-09T17:41:56.980 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:56.980 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:56.980 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:56.980 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:56 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:56.984 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-09T17:41:56.984 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph config generate-minimal-conf 2026-03-09T17:41:57.208 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:41:57.445 INFO:teuthology.orchestra.run.vm01.stdout:# minimal ceph.conf for 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:57.445 INFO:teuthology.orchestra.run.vm01.stdout:[global] 2026-03-09T17:41:57.445 INFO:teuthology.orchestra.run.vm01.stdout: fsid = 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:41:57.445 INFO:teuthology.orchestra.run.vm01.stdout: mon_host = [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] 2026-03-09T17:41:57.535 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-09T17:41:57.535 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:41:57.535 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T17:41:57.566 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:41:57.566 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:41:57.636 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:41:57.637 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T17:41:57.668 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:41:57.668 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:41:57.736 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-09T17:41:57.736 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T17:41:57.773 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-09T17:41:57.773 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: Updating vm01:/etc/ceph/ceph.conf 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: Updating vm05:/etc/ceph/ceph.conf 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: Updating vm01:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.conf 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: Updating vm05:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.conf 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: Updating vm04:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.conf 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: Reconfiguring mon.a (unknown last config time)... 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: Reconfiguring daemon mon.a on vm01 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/2394440900' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1435121758' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:57.795 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:57 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:57.838 INFO:tasks.cephadm:Adding mgr.x on vm01 2026-03-09T17:41:57.838 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch apply mgr '1;vm01=x' 2026-03-09T17:41:58.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: Updating vm01:/etc/ceph/ceph.conf 2026-03-09T17:41:58.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T17:41:58.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: Updating vm05:/etc/ceph/ceph.conf 2026-03-09T17:41:58.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: Updating vm01:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.conf 2026-03-09T17:41:58.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: Updating vm05:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.conf 2026-03-09T17:41:58.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: Updating vm04:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.conf 2026-03-09T17:41:58.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: Reconfiguring mon.a (unknown last config time)... 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: Reconfiguring daemon mon.a on vm01 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/2394440900' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1435121758' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:58.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:57 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.052 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.c/config 2026-03-09T17:41:58.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: Updating vm01:/etc/ceph/ceph.conf 2026-03-09T17:41:58.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: Updating vm04:/etc/ceph/ceph.conf 2026-03-09T17:41:58.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: Updating vm05:/etc/ceph/ceph.conf 2026-03-09T17:41:58.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: Updating vm01:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.conf 2026-03-09T17:41:58.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: Updating vm05:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.conf 2026-03-09T17:41:58.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: Updating vm04:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/config/ceph.conf 2026-03-09T17:41:58.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: Reconfiguring mon.a (unknown last config time)... 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: Reconfiguring daemon mon.a on vm01 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/2394440900' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1435121758' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:58.072 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:57 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.297 INFO:teuthology.orchestra.run.vm05.stdout:Scheduled mgr update... 2026-03-09T17:41:58.352 INFO:tasks.cephadm:Deploying OSDs... 2026-03-09T17:41:58.352 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:41:58.352 DEBUG:teuthology.orchestra.run.vm01:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T17:41:58.372 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:41:58.372 DEBUG:teuthology.orchestra.run.vm01:> ls /dev/[sv]d? 2026-03-09T17:41:58.429 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vda 2026-03-09T17:41:58.430 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdb 2026-03-09T17:41:58.430 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdc 2026-03-09T17:41:58.430 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdd 2026-03-09T17:41:58.430 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vde 2026-03-09T17:41:58.430 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T17:41:58.430 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T17:41:58.430 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdb 2026-03-09T17:41:58.489 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdb 2026-03-09T17:41:58.489 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:41:58.489 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 221 Links: 1 Device type: fc,10 2026-03-09T17:41:58.489 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:41:58.489 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:41:58.489 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-09 17:41:35.826464037 +0000 2026-03-09T17:41:58.489 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-09 17:34:53.219000000 +0000 2026-03-09T17:41:58.489 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-09 17:34:53.219000000 +0000 2026-03-09T17:41:58.489 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-09 17:34:51.234000000 +0000 2026-03-09T17:41:58.489 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T17:41:58.587 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-09T17:41:58.587 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-09T17:41:58.587 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000210685 s, 2.4 MB/s 2026-03-09T17:41:58.588 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T17:41:58.633 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdc 2026-03-09T17:41:58.659 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdc 2026-03-09T17:41:58.660 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:41:58.660 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 224 Links: 1 Device type: fc,20 2026-03-09T17:41:58.660 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:41:58.660 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:41:58.660 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-09 17:41:35.873464116 +0000 2026-03-09T17:41:58.660 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-09 17:34:53.203000000 +0000 2026-03-09T17:41:58.660 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-09 17:34:53.203000000 +0000 2026-03-09T17:41:58.660 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-09 17:34:51.246000000 +0000 2026-03-09T17:41:58.660 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T17:41:58.726 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-09T17:41:58.726 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-09T17:41:58.726 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000150762 s, 3.4 MB/s 2026-03-09T17:41:58.727 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T17:41:58.783 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdd 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: Reconfiguring mon.b (monmap changed)... 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: Reconfiguring daemon mon.b on vm04 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: Reconfiguring mon.c (monmap changed)... 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: Reconfiguring daemon mon.c on vm05 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:58 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.796 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:41:58 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x[53003]: 2026-03-09T17:41:58.494+0000 7f75e7f31640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: Reconfiguring mon.b (monmap changed)... 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: Reconfiguring daemon mon.b on vm04 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: Reconfiguring mon.c (monmap changed)... 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: Reconfiguring daemon mon.c on vm05 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.797 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:58 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:58.813 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdd 2026-03-09T17:41:58.813 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:41:58.813 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T17:41:58.813 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:41:58.813 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:41:58.813 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-09 17:41:35.916464188 +0000 2026-03-09T17:41:58.813 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-09 17:34:53.194000000 +0000 2026-03-09T17:41:58.813 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-09 17:34:53.194000000 +0000 2026-03-09T17:41:58.813 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-09 17:34:51.261000000 +0000 2026-03-09T17:41:58.813 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T17:41:58.876 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-09T17:41:58.876 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-09T17:41:58.876 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 9.5129e-05 s, 5.4 MB/s 2026-03-09T17:41:58.877 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T17:41:58.933 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vde 2026-03-09T17:41:58.991 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vde 2026-03-09T17:41:58.991 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:41:58.991 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T17:41:58.991 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:41:58.991 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:41:58.991 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-09 17:41:35.949464244 +0000 2026-03-09T17:41:58.991 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-09 17:34:53.236000000 +0000 2026-03-09T17:41:58.991 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-09 17:34:53.236000000 +0000 2026-03-09T17:41:58.991 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-09 17:34:51.267000000 +0000 2026-03-09T17:41:58.991 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T17:41:59.056 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-09T17:41:59.056 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-09T17:41:59.056 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000114103 s, 4.5 MB/s 2026-03-09T17:41:59.057 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T17:41:59.116 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:41:59.116 DEBUG:teuthology.orchestra.run.vm04:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T17:41:59.133 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:41:59.133 DEBUG:teuthology.orchestra.run.vm04:> ls /dev/[sv]d? 2026-03-09T17:41:59.189 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vda 2026-03-09T17:41:59.189 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdb 2026-03-09T17:41:59.189 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdc 2026-03-09T17:41:59.189 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdd 2026-03-09T17:41:59.189 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vde 2026-03-09T17:41:59.189 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T17:41:59.189 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T17:41:59.189 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdb 2026-03-09T17:41:59.246 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdb 2026-03-09T17:41:59.247 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:41:59.247 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-09T17:41:59.247 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:41:59.247 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:41:59.247 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-09 17:41:40.113299928 +0000 2026-03-09T17:41:59.247 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-09 17:36:32.298000000 +0000 2026-03-09T17:41:59.247 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-09 17:36:32.298000000 +0000 2026-03-09T17:41:59.247 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-09 17:36:30.272000000 +0000 2026-03-09T17:41:59.247 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: Reconfiguring mon.b (monmap changed)... 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: Reconfiguring daemon mon.b on vm04 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: Reconfiguring mon.c (monmap changed)... 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: Reconfiguring daemon mon.c on vm05 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:58 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:41:59.311 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-09T17:41:59.311 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-09T17:41:59.311 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000156755 s, 3.3 MB/s 2026-03-09T17:41:59.312 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T17:41:59.369 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdc 2026-03-09T17:41:59.427 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdc 2026-03-09T17:41:59.427 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:41:59.427 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-09T17:41:59.427 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:41:59.427 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:41:59.427 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-09 17:41:40.144300072 +0000 2026-03-09T17:41:59.427 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-09 17:36:32.300000000 +0000 2026-03-09T17:41:59.427 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-09 17:36:32.300000000 +0000 2026-03-09T17:41:59.427 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-09 17:36:30.277000000 +0000 2026-03-09T17:41:59.427 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T17:41:59.491 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-09T17:41:59.491 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-09T17:41:59.491 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000132719 s, 3.9 MB/s 2026-03-09T17:41:59.493 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T17:41:59.550 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdd 2026-03-09T17:41:59.610 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdd 2026-03-09T17:41:59.611 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:41:59.611 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T17:41:59.611 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:41:59.611 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:41:59.611 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-09 17:41:40.173300206 +0000 2026-03-09T17:41:59.611 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-09 17:36:32.303000000 +0000 2026-03-09T17:41:59.611 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-09 17:36:32.303000000 +0000 2026-03-09T17:41:59.611 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-09 17:36:30.288000000 +0000 2026-03-09T17:41:59.611 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T17:41:59.678 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-09T17:41:59.678 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-09T17:41:59.678 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000188073 s, 2.7 MB/s 2026-03-09T17:41:59.679 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T17:41:59.738 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vde 2026-03-09T17:41:59.802 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vde 2026-03-09T17:41:59.803 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:41:59.803 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T17:41:59.803 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:41:59.803 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:41:59.803 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-09 17:41:40.202300340 +0000 2026-03-09T17:41:59.803 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-09 17:36:32.305000000 +0000 2026-03-09T17:41:59.803 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-09 17:36:32.305000000 +0000 2026-03-09T17:41:59.803 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-09 17:36:30.295000000 +0000 2026-03-09T17:41:59.803 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T17:41:59.870 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-09T17:41:59.870 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-09T17:41:59.870 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000186478 s, 2.7 MB/s 2026-03-09T17:41:59.871 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T17:41:59.930 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-09T17:41:59.930 DEBUG:teuthology.orchestra.run.vm05:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T17:41:59.944 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:41:59.944 DEBUG:teuthology.orchestra.run.vm05:> ls /dev/[sv]d? 2026-03-09T17:42:00.000 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vda 2026-03-09T17:42:00.000 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vdb 2026-03-09T17:42:00.000 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vdc 2026-03-09T17:42:00.000 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vdd 2026-03-09T17:42:00.000 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vde 2026-03-09T17:42:00.001 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T17:42:00.001 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T17:42:00.001 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vdb 2026-03-09T17:42:00.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:59 vm04 ceph-mon[53158]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "1;vm01=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:00.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:59 vm04 ceph-mon[53158]: Saving service mgr spec with placement vm01=x;count:1 2026-03-09T17:42:00.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:59 vm04 ceph-mon[53158]: Reconfiguring mgr.x (unknown last config time)... 2026-03-09T17:42:00.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:41:59 vm04 ceph-mon[53158]: Reconfiguring daemon mgr.x on vm01 2026-03-09T17:42:00.058 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vdb 2026-03-09T17:42:00.058 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:42:00.058 INFO:teuthology.orchestra.run.vm05.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-09T17:42:00.058 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:42:00.058 INFO:teuthology.orchestra.run.vm05.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:42:00.058 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-09 17:41:43.793623507 +0000 2026-03-09T17:42:00.058 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-09 17:35:37.086000000 +0000 2026-03-09T17:42:00.058 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-09 17:35:37.086000000 +0000 2026-03-09T17:42:00.058 INFO:teuthology.orchestra.run.vm05.stdout: Birth: 2026-03-09 17:35:35.244000000 +0000 2026-03-09T17:42:00.058 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T17:42:00.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:59 vm05 ceph-mon[53831]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "1;vm01=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:00.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:59 vm05 ceph-mon[53831]: Saving service mgr spec with placement vm01=x;count:1 2026-03-09T17:42:00.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:59 vm05 ceph-mon[53831]: Reconfiguring mgr.x (unknown last config time)... 2026-03-09T17:42:00.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:41:59 vm05 ceph-mon[53831]: Reconfiguring daemon mgr.x on vm01 2026-03-09T17:42:00.094 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-09T17:42:00.094 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-09T17:42:00.094 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000205585 s, 2.5 MB/s 2026-03-09T17:42:00.095 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T17:42:00.156 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vdc 2026-03-09T17:42:00.216 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vdc 2026-03-09T17:42:00.216 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:42:00.216 INFO:teuthology.orchestra.run.vm05.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-09T17:42:00.216 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:42:00.216 INFO:teuthology.orchestra.run.vm05.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:42:00.216 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-09 17:41:43.831623616 +0000 2026-03-09T17:42:00.216 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-09 17:35:37.092000000 +0000 2026-03-09T17:42:00.216 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-09 17:35:37.092000000 +0000 2026-03-09T17:42:00.216 INFO:teuthology.orchestra.run.vm05.stdout: Birth: 2026-03-09 17:35:35.249000000 +0000 2026-03-09T17:42:00.216 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T17:42:00.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:59 vm01 ceph-mon[52793]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "1;vm01=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:00.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:59 vm01 ceph-mon[52793]: Saving service mgr spec with placement vm01=x;count:1 2026-03-09T17:42:00.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:59 vm01 ceph-mon[52793]: Reconfiguring mgr.x (unknown last config time)... 2026-03-09T17:42:00.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:41:59 vm01 ceph-mon[52793]: Reconfiguring daemon mgr.x on vm01 2026-03-09T17:42:00.282 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-09T17:42:00.282 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-09T17:42:00.282 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000216947 s, 2.4 MB/s 2026-03-09T17:42:00.283 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T17:42:00.343 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vdd 2026-03-09T17:42:00.403 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vdd 2026-03-09T17:42:00.403 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:42:00.403 INFO:teuthology.orchestra.run.vm05.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T17:42:00.403 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:42:00.403 INFO:teuthology.orchestra.run.vm05.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:42:00.403 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-09 17:41:43.867623719 +0000 2026-03-09T17:42:00.403 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-09 17:35:37.108000000 +0000 2026-03-09T17:42:00.403 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-09 17:35:37.108000000 +0000 2026-03-09T17:42:00.403 INFO:teuthology.orchestra.run.vm05.stdout: Birth: 2026-03-09 17:35:35.254000000 +0000 2026-03-09T17:42:00.403 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T17:42:00.468 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-09T17:42:00.468 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-09T17:42:00.468 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000254917 s, 2.0 MB/s 2026-03-09T17:42:00.469 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T17:42:00.529 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vde 2026-03-09T17:42:00.588 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vde 2026-03-09T17:42:00.588 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:42:00.588 INFO:teuthology.orchestra.run.vm05.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T17:42:00.588 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:42:00.588 INFO:teuthology.orchestra.run.vm05.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:42:00.588 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-09 17:41:43.893623794 +0000 2026-03-09T17:42:00.588 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-09 17:35:37.109000000 +0000 2026-03-09T17:42:00.588 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-09 17:35:37.109000000 +0000 2026-03-09T17:42:00.588 INFO:teuthology.orchestra.run.vm05.stdout: Birth: 2026-03-09 17:35:35.258000000 +0000 2026-03-09T17:42:00.588 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T17:42:00.650 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-09T17:42:00.650 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-09T17:42:00.650 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000161492 s, 3.2 MB/s 2026-03-09T17:42:00.651 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T17:42:00.706 INFO:tasks.cephadm:Deploying osd.0 on vm01 with /dev/vde... 2026-03-09T17:42:00.706 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- lvm zap /dev/vde 2026-03-09T17:42:00.866 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:42:00.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:00 vm01 ceph-mon[52793]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:01.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:00 vm05 ceph-mon[53831]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:01.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:00 vm04 ceph-mon[53158]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:01.659 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:42:01.676 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch daemon add osd vm01:/dev/vde 2026-03-09T17:42:01.838 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:42:02.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:02 vm01 ceph-mon[52793]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:02.924 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:02 vm01 ceph-mon[52793]: from='client.24110 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:02.925 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:02 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:02.925 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:02 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:02.925 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:02 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:03.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:02 vm05 ceph-mon[53831]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:03.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:02 vm05 ceph-mon[53831]: from='client.24110 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:03.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:02 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:03.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:02 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:03.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:02 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:03.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:02 vm04 ceph-mon[53158]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:03.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:02 vm04 ceph-mon[53158]: from='client.24110 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:03.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:02 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:03.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:02 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:03.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:02 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:04.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:03 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3037543539' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b4de0462-8141-43d0-a2b5-00b4ec8c42bb"}]: dispatch 2026-03-09T17:42:04.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:03 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3037543539' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "b4de0462-8141-43d0-a2b5-00b4ec8c42bb"}]': finished 2026-03-09T17:42:04.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:03 vm05 ceph-mon[53831]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T17:42:04.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:03 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:04.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:03 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2642898605' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:04.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:03 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3037543539' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b4de0462-8141-43d0-a2b5-00b4ec8c42bb"}]: dispatch 2026-03-09T17:42:04.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:03 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3037543539' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "b4de0462-8141-43d0-a2b5-00b4ec8c42bb"}]': finished 2026-03-09T17:42:04.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:03 vm01 ceph-mon[52793]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T17:42:04.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:03 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:04.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:03 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2642898605' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:04.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:03 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3037543539' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b4de0462-8141-43d0-a2b5-00b4ec8c42bb"}]: dispatch 2026-03-09T17:42:04.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:03 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3037543539' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "b4de0462-8141-43d0-a2b5-00b4ec8c42bb"}]': finished 2026-03-09T17:42:04.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:03 vm04 ceph-mon[53158]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T17:42:04.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:03 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:04.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:03 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2642898605' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:05.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:04 vm01 ceph-mon[52793]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:05.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:04 vm04 ceph-mon[53158]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:05.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:04 vm05 ceph-mon[53831]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:06.929 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:06 vm01 ceph-mon[52793]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:06.929 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:06 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T17:42:06.929 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:06 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:07.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:06 vm04 ceph-mon[53158]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:07.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:06 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T17:42:07.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:06 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:07.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:06 vm05 ceph-mon[53831]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:07.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:06 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T17:42:07.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:06 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:08.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:07 vm05 ceph-mon[53831]: Deploying daemon osd.0 on vm01 2026-03-09T17:42:08.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:07 vm01 ceph-mon[52793]: Deploying daemon osd.0 on vm01 2026-03-09T17:42:08.290 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:07 vm04 ceph-mon[53158]: Deploying daemon osd.0 on vm01 2026-03-09T17:42:08.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:08 vm01 ceph-mon[52793]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:08.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:08 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:08.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:08 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:08.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:08 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:09.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:08 vm04 ceph-mon[53158]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:09.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:08 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:09.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:08 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:09.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:08 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:09.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:08 vm05 ceph-mon[53831]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:09.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:08 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:09.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:08 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:09.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:08 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:09.926 INFO:teuthology.orchestra.run.vm01.stdout:Created osd(s) 0 on host 'vm01' 2026-03-09T17:42:09.998 DEBUG:teuthology.orchestra.run.vm01:osd.0> sudo journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.0.service 2026-03-09T17:42:09.999 INFO:tasks.cephadm:Deploying osd.1 on vm01 with /dev/vdd... 2026-03-09T17:42:09.999 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- lvm zap /dev/vdd 2026-03-09T17:42:10.275 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:42:10.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:10 vm04 ceph-mon[53158]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:10.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:10 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:10.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:10 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:10.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:10 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:10.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:10 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:10.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:10 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:10.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:10 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:10.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:10 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:10.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:10 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:10.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:10 vm05 ceph-mon[53831]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:10.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:10 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:10.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:10 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:10.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:10 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:10.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:10 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:10.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:10 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:10.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:10 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:10.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:10 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:10.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:10 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:10.844 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:42:10 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-0[64363]: 2026-03-09T17:42:10.717+0000 7fe10549c740 -1 osd.0 0 log_to_monitors true 2026-03-09T17:42:10.844 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:10 vm01 ceph-mon[52793]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:10.844 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:10 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:10.844 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:10 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:10.844 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:10 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:10.844 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:10 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:10.844 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:10 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:10.844 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:10 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:10.844 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:10 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:10.844 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:10 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:11.640 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:42:11.658 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch daemon add osd vm01:/dev/vdd 2026-03-09T17:42:11.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:11 vm04 ceph-mon[53158]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T17:42:11.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:11 vm04 ceph-mon[53158]: Detected new or changed devices on vm01 2026-03-09T17:42:11.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:11 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:11.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:11 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:11.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:11 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:11.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:11 vm04 ceph-mon[53158]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-09T17:42:11.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:11 vm04 ceph-mon[53158]: Unable to set osd_memory_target on vm01 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T17:42:11.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:11 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:11.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:11 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:11.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:11 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:11.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:11 vm05 ceph-mon[53831]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T17:42:11.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:11 vm05 ceph-mon[53831]: Detected new or changed devices on vm01 2026-03-09T17:42:11.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:11 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:11.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:11 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:11.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:11 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:11.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:11 vm05 ceph-mon[53831]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-09T17:42:11.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:11 vm05 ceph-mon[53831]: Unable to set osd_memory_target on vm01 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T17:42:11.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:11 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:11.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:11 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:11.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:11 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:11.829 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:42:11.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:11 vm01 ceph-mon[52793]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T17:42:11.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:11 vm01 ceph-mon[52793]: Detected new or changed devices on vm01 2026-03-09T17:42:11.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:11 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:11.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:11 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:11.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:11 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:11.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:11 vm01 ceph-mon[52793]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-09T17:42:11.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:11 vm01 ceph-mon[52793]: Unable to set osd_memory_target on vm01 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T17:42:11.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:11 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:11.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:11 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:11.856 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:11 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:12.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:12 vm01 ceph-mon[52793]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:12.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:12 vm01 ceph-mon[52793]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T17:42:12.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:12 vm01 ceph-mon[52793]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T17:42:12.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:12 vm01 ceph-mon[52793]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-09T17:42:12.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:12 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:12.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:12 vm01 ceph-mon[52793]: from='client.14238 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:12.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:12 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:12.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:12 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:12.692 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:12 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:12.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:12 vm04 ceph-mon[53158]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:12.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:12 vm04 ceph-mon[53158]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T17:42:12.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:12 vm04 ceph-mon[53158]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T17:42:12.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:12 vm04 ceph-mon[53158]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-09T17:42:12.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:12 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:12.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:12 vm04 ceph-mon[53158]: from='client.14238 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:12.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:12 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:12.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:12 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:12.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:12 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:12.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:12 vm05 ceph-mon[53831]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:12.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:12 vm05 ceph-mon[53831]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T17:42:12.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:12 vm05 ceph-mon[53831]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T17:42:12.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:12 vm05 ceph-mon[53831]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-09T17:42:12.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:12 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:12.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:12 vm05 ceph-mon[53831]: from='client.14238 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:12.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:12 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:12.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:12 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:12.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:12 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:13 vm01 ceph-mon[52793]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-09T17:42:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:13 vm01 ceph-mon[52793]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T17:42:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:13 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:13 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:13 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2795976069' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "651f912d-fa94-4d83-9b72-bb0b98168d7b"}]: dispatch 2026-03-09T17:42:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:13 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2795976069' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "651f912d-fa94-4d83-9b72-bb0b98168d7b"}]': finished 2026-03-09T17:42:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:13 vm01 ceph-mon[52793]: osdmap e8: 2 total, 0 up, 2 in 2026-03-09T17:42:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:13 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:13 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:13 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/997199398' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:13 vm01 ceph-mon[52793]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' 2026-03-09T17:42:13.728 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:42:13 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-0[64363]: 2026-03-09T17:42:13.419+0000 7fe10141d640 -1 osd.0 0 waiting for initial osdmap 2026-03-09T17:42:13.728 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:42:13 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-0[64363]: 2026-03-09T17:42:13.427+0000 7fe0fca46640 -1 osd.0 8 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:42:13.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:13 vm04 ceph-mon[53158]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-09T17:42:13.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:13 vm04 ceph-mon[53158]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T17:42:13.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:13 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:13.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:13 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:13.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:13 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2795976069' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "651f912d-fa94-4d83-9b72-bb0b98168d7b"}]: dispatch 2026-03-09T17:42:13.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:13 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2795976069' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "651f912d-fa94-4d83-9b72-bb0b98168d7b"}]': finished 2026-03-09T17:42:13.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:13 vm04 ceph-mon[53158]: osdmap e8: 2 total, 0 up, 2 in 2026-03-09T17:42:13.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:13 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:13.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:13 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:13.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:13 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/997199398' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:13.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:13 vm04 ceph-mon[53158]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' 2026-03-09T17:42:13.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:13 vm05 ceph-mon[53831]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-09T17:42:13.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:13 vm05 ceph-mon[53831]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T17:42:13.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:13 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:13.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:13 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:13.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:13 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2795976069' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "651f912d-fa94-4d83-9b72-bb0b98168d7b"}]: dispatch 2026-03-09T17:42:13.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:13 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2795976069' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "651f912d-fa94-4d83-9b72-bb0b98168d7b"}]': finished 2026-03-09T17:42:13.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:13 vm05 ceph-mon[53831]: osdmap e8: 2 total, 0 up, 2 in 2026-03-09T17:42:13.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:13 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:13.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:13 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:13.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:13 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/997199398' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:13.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:13 vm05 ceph-mon[53831]: from='osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347]' entity='osd.0' 2026-03-09T17:42:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:14 vm04 ceph-mon[53158]: purged_snaps scrub starts 2026-03-09T17:42:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:14 vm04 ceph-mon[53158]: purged_snaps scrub ok 2026-03-09T17:42:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:14 vm04 ceph-mon[53158]: pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:14 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:14 vm04 ceph-mon[53158]: osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347] boot 2026-03-09T17:42:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:14 vm04 ceph-mon[53158]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T17:42:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:14 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:14 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:14.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:14 vm05 ceph-mon[53831]: purged_snaps scrub starts 2026-03-09T17:42:14.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:14 vm05 ceph-mon[53831]: purged_snaps scrub ok 2026-03-09T17:42:14.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:14 vm05 ceph-mon[53831]: pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:14.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:14 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:14.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:14 vm05 ceph-mon[53831]: osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347] boot 2026-03-09T17:42:14.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:14 vm05 ceph-mon[53831]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T17:42:14.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:14 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:14.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:14 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:14.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:14 vm01 ceph-mon[52793]: purged_snaps scrub starts 2026-03-09T17:42:14.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:14 vm01 ceph-mon[52793]: purged_snaps scrub ok 2026-03-09T17:42:14.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:14 vm01 ceph-mon[52793]: pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:42:14.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:14 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:14.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:14 vm01 ceph-mon[52793]: osd.0 [v2:192.168.123.101:6802/3943236347,v1:192.168.123.101:6803/3943236347] boot 2026-03-09T17:42:14.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:14 vm01 ceph-mon[52793]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T17:42:14.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:14 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:42:14.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:14 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:16.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:16 vm01 ceph-mon[52793]: pgmap v21: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:16.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:16 vm01 ceph-mon[52793]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T17:42:16.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:16 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:16.790 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:16 vm04 ceph-mon[53158]: pgmap v21: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:16.790 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:16 vm04 ceph-mon[53158]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T17:42:16.790 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:16 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:16.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:16 vm05 ceph-mon[53831]: pgmap v21: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:16.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:16 vm05 ceph-mon[53831]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T17:42:16.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:16 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:17.605 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:17 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T17:42:17.605 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:17 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:17.605 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:17 vm01 ceph-mon[52793]: Deploying daemon osd.1 on vm01 2026-03-09T17:42:17.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:17 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T17:42:17.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:17 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:17.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:17 vm05 ceph-mon[53831]: Deploying daemon osd.1 on vm01 2026-03-09T17:42:18.040 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:17 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T17:42:18.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:17 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:18.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:17 vm04 ceph-mon[53158]: Deploying daemon osd.1 on vm01 2026-03-09T17:42:18.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:18 vm01 ceph-mon[52793]: pgmap v23: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:18.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:18 vm05 ceph-mon[53831]: pgmap v23: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:19.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:18 vm04 ceph-mon[53158]: pgmap v23: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:19.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:19 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:19.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:19 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:19.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:19 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:19.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:19 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:19.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:19 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:19.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:19 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:20.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:19 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:20.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:19 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:20.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:19 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:20.138 INFO:teuthology.orchestra.run.vm01.stdout:Created osd(s) 1 on host 'vm01' 2026-03-09T17:42:20.200 DEBUG:teuthology.orchestra.run.vm01:osd.1> sudo journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.1.service 2026-03-09T17:42:20.202 INFO:tasks.cephadm:Deploying osd.2 on vm04 with /dev/vde... 2026-03-09T17:42:20.202 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- lvm zap /dev/vde 2026-03-09T17:42:20.354 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:42:20 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-1[69215]: 2026-03-09T17:42:20.351+0000 7fc08dfd4740 -1 osd.1 0 log_to_monitors true 2026-03-09T17:42:20.408 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.b/config 2026-03-09T17:42:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:20 vm04 ceph-mon[53158]: pgmap v24: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:20 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:20 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:20 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:20 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:20 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:20 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:20 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:20 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:20 vm04 ceph-mon[53158]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T17:42:20.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:20 vm01 ceph-mon[52793]: pgmap v24: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:20.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:20 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:20.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:20 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:20.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:20 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:20.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:20 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:20.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:20 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:20.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:20 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:20.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:20 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:20.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:20 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:20.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:20 vm01 ceph-mon[52793]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T17:42:21.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:20 vm05 ceph-mon[53831]: pgmap v24: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:21.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:20 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:21.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:20 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:21.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:20 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:21.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:20 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:21.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:20 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:21.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:20 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:21.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:20 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:21.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:20 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:21.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:20 vm05 ceph-mon[53831]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T17:42:21.372 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:42:21.399 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch daemon add osd vm04:/dev/vde 2026-03-09T17:42:21.587 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.b/config 2026-03-09T17:42:22.156 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T17:42:22.156 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T17:42:22.156 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: Detected new or changed devices on vm01 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: Adjusting osd_memory_target on vm01 to 128.5M 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: Unable to set osd_memory_target on vm01 to 134765363: error parsing value: Value '134765363' is below minimum 939524096 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: from='client.24146 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:22.160 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:22 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:22.465 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T17:42:22.465 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: Detected new or changed devices on vm01 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: Adjusting osd_memory_target on vm01 to 128.5M 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: Unable to set osd_memory_target on vm01 to 134765363: error parsing value: Value '134765363' is below minimum 939524096 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: from='client.24146 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:22.466 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:22 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:22.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T17:42:22.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T17:42:22.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: Detected new or changed devices on vm01 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: Adjusting osd_memory_target on vm01 to 128.5M 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: Unable to set osd_memory_target on vm01 to 134765363: error parsing value: Value '134765363' is below minimum 939524096 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: from='client.24146 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:22.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:22 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:23.478 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:42:23 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-1[69215]: 2026-03-09T17:42:23.039+0000 7fc08a768640 -1 osd.1 0 waiting for initial osdmap 2026-03-09T17:42:23.478 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:42:23 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-1[69215]: 2026-03-09T17:42:23.046+0000 7fc08557e640 -1 osd.1 13 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:42:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:23 vm01 ceph-mon[52793]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-09T17:42:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:23 vm01 ceph-mon[52793]: osdmap e12: 2 total, 1 up, 2 in 2026-03-09T17:42:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:23 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:23 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:23 vm01 ceph-mon[52793]: from='client.? 192.168.123.104:0/2574372256' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9afeb5af-65ac-471d-b3a4-b633e11482f9"}]: dispatch 2026-03-09T17:42:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:23 vm01 ceph-mon[52793]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9afeb5af-65ac-471d-b3a4-b633e11482f9"}]: dispatch 2026-03-09T17:42:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:23 vm01 ceph-mon[52793]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "9afeb5af-65ac-471d-b3a4-b633e11482f9"}]': finished 2026-03-09T17:42:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:23 vm01 ceph-mon[52793]: osdmap e13: 3 total, 1 up, 3 in 2026-03-09T17:42:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:23 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:23 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:23 vm01 ceph-mon[52793]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' 2026-03-09T17:42:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:23 vm01 ceph-mon[52793]: from='client.? 192.168.123.104:0/3256159029' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:23.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:23 vm04 ceph-mon[53158]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-09T17:42:23.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:23 vm04 ceph-mon[53158]: osdmap e12: 2 total, 1 up, 2 in 2026-03-09T17:42:23.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:23 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:23.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:23 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:23.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:23 vm04 ceph-mon[53158]: from='client.? 192.168.123.104:0/2574372256' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9afeb5af-65ac-471d-b3a4-b633e11482f9"}]: dispatch 2026-03-09T17:42:23.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:23 vm04 ceph-mon[53158]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9afeb5af-65ac-471d-b3a4-b633e11482f9"}]: dispatch 2026-03-09T17:42:23.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:23 vm04 ceph-mon[53158]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "9afeb5af-65ac-471d-b3a4-b633e11482f9"}]': finished 2026-03-09T17:42:23.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:23 vm04 ceph-mon[53158]: osdmap e13: 3 total, 1 up, 3 in 2026-03-09T17:42:23.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:23 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:23.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:23 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:23.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:23 vm04 ceph-mon[53158]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' 2026-03-09T17:42:23.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:23 vm04 ceph-mon[53158]: from='client.? 192.168.123.104:0/3256159029' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:23.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:23 vm05 ceph-mon[53831]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-09T17:42:23.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:23 vm05 ceph-mon[53831]: osdmap e12: 2 total, 1 up, 2 in 2026-03-09T17:42:23.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:23 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:23.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:23 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:23.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:23 vm05 ceph-mon[53831]: from='client.? 192.168.123.104:0/2574372256' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9afeb5af-65ac-471d-b3a4-b633e11482f9"}]: dispatch 2026-03-09T17:42:23.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:23 vm05 ceph-mon[53831]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9afeb5af-65ac-471d-b3a4-b633e11482f9"}]: dispatch 2026-03-09T17:42:23.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:23 vm05 ceph-mon[53831]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "9afeb5af-65ac-471d-b3a4-b633e11482f9"}]': finished 2026-03-09T17:42:23.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:23 vm05 ceph-mon[53831]: osdmap e13: 3 total, 1 up, 3 in 2026-03-09T17:42:23.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:23 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:23.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:23 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:23.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:23 vm05 ceph-mon[53831]: from='osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663]' entity='osd.1' 2026-03-09T17:42:23.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:23 vm05 ceph-mon[53831]: from='client.? 192.168.123.104:0/3256159029' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:24.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:24 vm01 ceph-mon[52793]: purged_snaps scrub starts 2026-03-09T17:42:24.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:24 vm01 ceph-mon[52793]: purged_snaps scrub ok 2026-03-09T17:42:24.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:24 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:24.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:24 vm01 ceph-mon[52793]: pgmap v29: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:24.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:24 vm01 ceph-mon[52793]: osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663] boot 2026-03-09T17:42:24.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:24 vm01 ceph-mon[52793]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T17:42:24.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:24 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:24.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:24 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:24.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:24 vm04 ceph-mon[53158]: purged_snaps scrub starts 2026-03-09T17:42:24.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:24 vm04 ceph-mon[53158]: purged_snaps scrub ok 2026-03-09T17:42:24.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:24 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:24.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:24 vm04 ceph-mon[53158]: pgmap v29: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:24.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:24 vm04 ceph-mon[53158]: osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663] boot 2026-03-09T17:42:24.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:24 vm04 ceph-mon[53158]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T17:42:24.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:24 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:24.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:24 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:24.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:24 vm05 ceph-mon[53831]: purged_snaps scrub starts 2026-03-09T17:42:24.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:24 vm05 ceph-mon[53831]: purged_snaps scrub ok 2026-03-09T17:42:24.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:24 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:24.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:24 vm05 ceph-mon[53831]: pgmap v29: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:42:24.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:24 vm05 ceph-mon[53831]: osd.1 [v2:192.168.123.101:6810/2752039663,v1:192.168.123.101:6811/2752039663] boot 2026-03-09T17:42:24.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:24 vm05 ceph-mon[53831]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T17:42:24.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:24 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:42:24.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:24 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:26.173 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:26 vm04 ceph-mon[53158]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T17:42:26.173 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:26 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:26.173 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:26 vm04 ceph-mon[53158]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:26.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:26 vm01 ceph-mon[52793]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T17:42:26.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:26 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:26.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:26 vm01 ceph-mon[52793]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:26.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:26 vm05 ceph-mon[53831]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T17:42:26.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:26 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:26.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:26 vm05 ceph-mon[53831]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:27.475 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:27 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T17:42:27.475 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:27 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:27.475 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:27 vm04 ceph-mon[53158]: Deploying daemon osd.2 on vm04 2026-03-09T17:42:27.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:27 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T17:42:27.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:27 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:27.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:27 vm01 ceph-mon[52793]: Deploying daemon osd.2 on vm04 2026-03-09T17:42:27.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:27 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T17:42:27.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:27 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:27.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:27 vm05 ceph-mon[53831]: Deploying daemon osd.2 on vm04 2026-03-09T17:42:28.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:28 vm01 ceph-mon[52793]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:28.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:28 vm04 ceph-mon[53158]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:28.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:28 vm05 ceph-mon[53831]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:29.447 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:29 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:29.447 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:29 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:29.447 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:29 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:29.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:29 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:29.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:29 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:29.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:29 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:29.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:29 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:29.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:29 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:29.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:29 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.099 INFO:teuthology.orchestra.run.vm04.stdout:Created osd(s) 2 on host 'vm04' 2026-03-09T17:42:30.165 DEBUG:teuthology.orchestra.run.vm04:osd.2> sudo journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.2.service 2026-03-09T17:42:30.166 INFO:tasks.cephadm:Deploying osd.3 on vm04 with /dev/vdd... 2026-03-09T17:42:30.166 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- lvm zap /dev/vdd 2026-03-09T17:42:30.491 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.b/config 2026-03-09T17:42:30.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:30 vm04 ceph-mon[53158]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:30.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:30 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:30 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:30 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:30.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:30 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:30.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:30 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:30 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:30.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:30 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:30 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.791 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:42:30 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-2[56304]: 2026-03-09T17:42:30.625+0000 7f4036940740 -1 osd.2 0 log_to_monitors true 2026-03-09T17:42:30.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:30 vm05 ceph-mon[53831]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:30.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:30 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:30 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:30 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:30.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:30 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:30.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:30 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:30 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:30.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:30 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:30 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:30 vm01 ceph-mon[52793]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:30.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:30 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:30 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:30 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:30.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:30 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:30.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:30 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:30 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:30.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:30 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:30.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:30 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:31.576 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:31 vm04 ceph-mon[53158]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T17:42:31.576 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:31 vm04 ceph-mon[53158]: from='osd.2 [v2:192.168.123.104:6800/3952920925,v1:192.168.123.104:6801/3952920925]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T17:42:31.576 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:31 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:31.576 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:31 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:31.576 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:31 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:31.576 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:31 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:31.576 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:31 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:31.576 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:31 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:31.576 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:31 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:31.681 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:42:31.702 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch daemon add osd vm04:/dev/vdd 2026-03-09T17:42:31.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:31 vm05 ceph-mon[53831]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T17:42:31.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:31 vm05 ceph-mon[53831]: from='osd.2 [v2:192.168.123.104:6800/3952920925,v1:192.168.123.104:6801/3952920925]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T17:42:31.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:31 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:31.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:31 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:31.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:31 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:31.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:31 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:31.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:31 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:31.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:31 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:31.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:31 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:31.879 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.b/config 2026-03-09T17:42:31.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:31 vm01 ceph-mon[52793]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T17:42:31.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:31 vm01 ceph-mon[52793]: from='osd.2 [v2:192.168.123.104:6800/3952920925,v1:192.168.123.104:6801/3952920925]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T17:42:31.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:31.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:31.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:31.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:31.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:31.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:31.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:31 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:32.729 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:32 vm04 ceph-mon[53158]: Detected new or changed devices on vm04 2026-03-09T17:42:32.729 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:32 vm04 ceph-mon[53158]: Adjusting osd_memory_target on vm04 to 4353M 2026-03-09T17:42:32.729 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:32 vm04 ceph-mon[53158]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:32.729 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:32 vm04 ceph-mon[53158]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T17:42:32.729 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:32 vm04 ceph-mon[53158]: osdmap e16: 3 total, 2 up, 3 in 2026-03-09T17:42:32.729 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:32 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:32.729 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:32 vm04 ceph-mon[53158]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T17:42:32.729 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:32 vm04 ceph-mon[53158]: from='osd.2 [v2:192.168.123.104:6800/3952920925,v1:192.168.123.104:6801/3952920925]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T17:42:32.729 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:32 vm04 ceph-mon[53158]: from='client.24173 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:32.729 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:32 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:32.729 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:32 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:32.729 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:32 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:32.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:32 vm05 ceph-mon[53831]: Detected new or changed devices on vm04 2026-03-09T17:42:32.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:32 vm05 ceph-mon[53831]: Adjusting osd_memory_target on vm04 to 4353M 2026-03-09T17:42:32.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:32 vm05 ceph-mon[53831]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:32.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:32 vm05 ceph-mon[53831]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T17:42:32.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:32 vm05 ceph-mon[53831]: osdmap e16: 3 total, 2 up, 3 in 2026-03-09T17:42:32.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:32 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:32.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:32 vm05 ceph-mon[53831]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T17:42:32.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:32 vm05 ceph-mon[53831]: from='osd.2 [v2:192.168.123.104:6800/3952920925,v1:192.168.123.104:6801/3952920925]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T17:42:32.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:32 vm05 ceph-mon[53831]: from='client.24173 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:32.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:32 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:32.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:32 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:32.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:32 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:32.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:32 vm01 ceph-mon[52793]: Detected new or changed devices on vm04 2026-03-09T17:42:32.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:32 vm01 ceph-mon[52793]: Adjusting osd_memory_target on vm04 to 4353M 2026-03-09T17:42:32.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:32 vm01 ceph-mon[52793]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:32.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:32 vm01 ceph-mon[52793]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T17:42:32.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:32 vm01 ceph-mon[52793]: osdmap e16: 3 total, 2 up, 3 in 2026-03-09T17:42:32.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:32 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:32.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:32 vm01 ceph-mon[52793]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T17:42:32.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:32 vm01 ceph-mon[52793]: from='osd.2 [v2:192.168.123.104:6800/3952920925,v1:192.168.123.104:6801/3952920925]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T17:42:32.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:32 vm01 ceph-mon[52793]: from='client.24173 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:32.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:32 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:32.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:32 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:32.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:32 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:33.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:33 vm04 ceph-mon[53158]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T17:42:33.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:33 vm04 ceph-mon[53158]: osdmap e17: 3 total, 2 up, 3 in 2026-03-09T17:42:33.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:33 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:33.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:33 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:33.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:33 vm04 ceph-mon[53158]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "69f44bac-cbb9-4564-abdc-37247aa0c954"}]: dispatch 2026-03-09T17:42:33.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:33 vm04 ceph-mon[53158]: from='client.? 192.168.123.104:0/3296920030' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "69f44bac-cbb9-4564-abdc-37247aa0c954"}]: dispatch 2026-03-09T17:42:33.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:33 vm04 ceph-mon[53158]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "69f44bac-cbb9-4564-abdc-37247aa0c954"}]': finished 2026-03-09T17:42:33.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:33 vm04 ceph-mon[53158]: osdmap e18: 4 total, 2 up, 4 in 2026-03-09T17:42:33.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:33 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:33.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:33 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:33.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:33 vm04 ceph-mon[53158]: from='osd.2 ' entity='osd.2' 2026-03-09T17:42:33.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:33 vm04 ceph-mon[53158]: from='client.? 192.168.123.104:0/1955652912' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:33.791 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:42:33 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-2[56304]: 2026-03-09T17:42:33.454+0000 7f40328c1640 -1 osd.2 0 waiting for initial osdmap 2026-03-09T17:42:33.791 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:42:33 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-2[56304]: 2026-03-09T17:42:33.465+0000 7f402deea640 -1 osd.2 18 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:42:33.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:33 vm05 ceph-mon[53831]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T17:42:33.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:33 vm05 ceph-mon[53831]: osdmap e17: 3 total, 2 up, 3 in 2026-03-09T17:42:33.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:33 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:33.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:33 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:33.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:33 vm05 ceph-mon[53831]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "69f44bac-cbb9-4564-abdc-37247aa0c954"}]: dispatch 2026-03-09T17:42:33.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:33 vm05 ceph-mon[53831]: from='client.? 192.168.123.104:0/3296920030' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "69f44bac-cbb9-4564-abdc-37247aa0c954"}]: dispatch 2026-03-09T17:42:33.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:33 vm05 ceph-mon[53831]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "69f44bac-cbb9-4564-abdc-37247aa0c954"}]': finished 2026-03-09T17:42:33.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:33 vm05 ceph-mon[53831]: osdmap e18: 4 total, 2 up, 4 in 2026-03-09T17:42:33.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:33 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:33.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:33 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:33.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:33 vm05 ceph-mon[53831]: from='osd.2 ' entity='osd.2' 2026-03-09T17:42:33.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:33 vm05 ceph-mon[53831]: from='client.? 192.168.123.104:0/1955652912' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:33.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:33 vm01 ceph-mon[52793]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T17:42:33.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:33 vm01 ceph-mon[52793]: osdmap e17: 3 total, 2 up, 3 in 2026-03-09T17:42:33.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:33.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:33.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:33 vm01 ceph-mon[52793]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "69f44bac-cbb9-4564-abdc-37247aa0c954"}]: dispatch 2026-03-09T17:42:33.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:33 vm01 ceph-mon[52793]: from='client.? 192.168.123.104:0/3296920030' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "69f44bac-cbb9-4564-abdc-37247aa0c954"}]: dispatch 2026-03-09T17:42:33.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:33 vm01 ceph-mon[52793]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "69f44bac-cbb9-4564-abdc-37247aa0c954"}]': finished 2026-03-09T17:42:33.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:33 vm01 ceph-mon[52793]: osdmap e18: 4 total, 2 up, 4 in 2026-03-09T17:42:33.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:33.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:33.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:33 vm01 ceph-mon[52793]: from='osd.2 ' entity='osd.2' 2026-03-09T17:42:33.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:33 vm01 ceph-mon[52793]: from='client.? 192.168.123.104:0/1955652912' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:34.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:34 vm05 ceph-mon[53831]: purged_snaps scrub starts 2026-03-09T17:42:34.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:34 vm05 ceph-mon[53831]: purged_snaps scrub ok 2026-03-09T17:42:34.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:34 vm05 ceph-mon[53831]: pgmap v39: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:34.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:34 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:34.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:34 vm05 ceph-mon[53831]: osd.2 [v2:192.168.123.104:6800/3952920925,v1:192.168.123.104:6801/3952920925] boot 2026-03-09T17:42:34.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:34 vm05 ceph-mon[53831]: osdmap e19: 4 total, 3 up, 4 in 2026-03-09T17:42:34.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:34 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:34.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:34 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:34.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:34 vm01 ceph-mon[52793]: purged_snaps scrub starts 2026-03-09T17:42:34.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:34 vm01 ceph-mon[52793]: purged_snaps scrub ok 2026-03-09T17:42:34.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:34 vm01 ceph-mon[52793]: pgmap v39: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:34.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:34 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:34.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:34 vm01 ceph-mon[52793]: osd.2 [v2:192.168.123.104:6800/3952920925,v1:192.168.123.104:6801/3952920925] boot 2026-03-09T17:42:34.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:34 vm01 ceph-mon[52793]: osdmap e19: 4 total, 3 up, 4 in 2026-03-09T17:42:34.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:34 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:34.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:34 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:35.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:34 vm04 ceph-mon[53158]: purged_snaps scrub starts 2026-03-09T17:42:35.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:34 vm04 ceph-mon[53158]: purged_snaps scrub ok 2026-03-09T17:42:35.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:34 vm04 ceph-mon[53158]: pgmap v39: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:42:35.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:34 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:35.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:34 vm04 ceph-mon[53158]: osd.2 [v2:192.168.123.104:6800/3952920925,v1:192.168.123.104:6801/3952920925] boot 2026-03-09T17:42:35.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:34 vm04 ceph-mon[53158]: osdmap e19: 4 total, 3 up, 4 in 2026-03-09T17:42:35.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:34 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:42:35.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:34 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:35.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:35 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T17:42:36.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:35 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T17:42:36.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:35 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T17:42:36.731 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:36 vm04 ceph-mon[53158]: pgmap v41: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:36.731 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T17:42:36.731 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:36 vm04 ceph-mon[53158]: osdmap e20: 4 total, 3 up, 4 in 2026-03-09T17:42:36.731 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:36.731 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T17:42:36.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:36 vm01 ceph-mon[52793]: pgmap v41: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:36.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T17:42:36.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:36 vm01 ceph-mon[52793]: osdmap e20: 4 total, 3 up, 4 in 2026-03-09T17:42:36.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:36.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T17:42:37.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:36 vm05 ceph-mon[53831]: pgmap v41: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:37.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T17:42:37.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:36 vm05 ceph-mon[53831]: osdmap e20: 4 total, 3 up, 4 in 2026-03-09T17:42:37.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:37.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T17:42:37.660 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:37 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T17:42:37.660 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:37 vm04 ceph-mon[53158]: osdmap e21: 4 total, 3 up, 4 in 2026-03-09T17:42:37.660 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:37 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:37.660 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:37 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-09T17:42:37.660 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:37 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:37.660 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:42:37 vm04 sudo[60607]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-09T17:42:37.660 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:42:37 vm04 sudo[60607]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T17:42:37.660 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:42:37 vm04 sudo[60607]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T17:42:37.660 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:42:37 vm04 sudo[60607]: pam_unix(sudo:session): session closed for user root 2026-03-09T17:42:37.926 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:37 vm04 sudo[60619]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T17:42:37.926 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:37 vm04 sudo[60619]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T17:42:37.926 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:37 vm04 sudo[60619]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T17:42:37.926 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:37 vm04 sudo[60619]: pam_unix(sudo:session): session closed for user root 2026-03-09T17:42:37.978 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:42:37 vm01 sudo[72489]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vdd 2026-03-09T17:42:37.978 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:42:37 vm01 sudo[72489]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T17:42:37.978 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:42:37 vm01 sudo[72489]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T17:42:37.978 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:42:37 vm01 sudo[72489]: pam_unix(sudo:session): session closed for user root 2026-03-09T17:42:37.978 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:42:37 vm01 sudo[72485]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-09T17:42:37.978 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:42:37 vm01 sudo[72485]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T17:42:37.978 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:42:37 vm01 sudo[72485]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T17:42:37.978 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:42:37 vm01 sudo[72485]: pam_unix(sudo:session): session closed for user root 2026-03-09T17:42:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T17:42:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:37 vm01 ceph-mon[52793]: osdmap e21: 4 total, 3 up, 4 in 2026-03-09T17:42:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-09T17:42:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:37 vm01 sudo[72493]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T17:42:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:37 vm01 sudo[72493]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T17:42:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:37 vm01 sudo[72493]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T17:42:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:37 vm01 sudo[72493]: pam_unix(sudo:session): session closed for user root 2026-03-09T17:42:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:37 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T17:42:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:37 vm05 ceph-mon[53831]: osdmap e21: 4 total, 3 up, 4 in 2026-03-09T17:42:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:37 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:37 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-09T17:42:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:37 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:37 vm05 sudo[55211]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T17:42:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:37 vm05 sudo[55211]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T17:42:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:37 vm05 sudo[55211]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T17:42:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:37 vm05 sudo[55211]: pam_unix(sudo:session): session closed for user root 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: pgmap v44: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: Deploying daemon osd.3 on vm04 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: osdmap e22: 4 total, 3 up, 4 in 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T17:42:38.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:38 vm01 ceph-mon[52793]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T17:42:39.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: pgmap v44: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:39.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: Deploying daemon osd.3 on vm04 2026-03-09T17:42:39.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: osdmap e22: 4 total, 3 up, 4 in 2026-03-09T17:42:39.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T17:42:39.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:38 vm04 ceph-mon[53158]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T17:42:39.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: pgmap v44: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:39.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: Deploying daemon osd.3 on vm04 2026-03-09T17:42:39.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: osdmap e22: 4 total, 3 up, 4 in 2026-03-09T17:42:39.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:39.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T17:42:39.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T17:42:39.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:42:39.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:42:39.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:42:39.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:42:39.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:42:39.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:42:39.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T17:42:39.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T17:42:39.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T17:42:39.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T17:42:39.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T17:42:39.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T17:42:39.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:38 vm05 ceph-mon[53831]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T17:42:40.866 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:40 vm04 ceph-mon[53158]: pgmap v46: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:40.866 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:40 vm04 ceph-mon[53158]: mgrmap e14: x(active, since 68s) 2026-03-09T17:42:40.866 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:40.866 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:40.866 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:40.866 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:40.866 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:40.939 INFO:teuthology.orchestra.run.vm04.stdout:Created osd(s) 3 on host 'vm04' 2026-03-09T17:42:40.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:40 vm01 ceph-mon[52793]: pgmap v46: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:40.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:40 vm01 ceph-mon[52793]: mgrmap e14: x(active, since 68s) 2026-03-09T17:42:40.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:40.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:40.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:40.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:40.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:40.999 DEBUG:teuthology.orchestra.run.vm04:osd.3> sudo journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.3.service 2026-03-09T17:42:41.000 INFO:tasks.cephadm:Deploying osd.4 on vm04 with /dev/vdc... 2026-03-09T17:42:41.000 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- lvm zap /dev/vdc 2026-03-09T17:42:41.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:40 vm05 ceph-mon[53831]: pgmap v46: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:41.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:40 vm05 ceph-mon[53831]: mgrmap e14: x(active, since 68s) 2026-03-09T17:42:41.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:41.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:41.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:41.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:41.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:41.294 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.b/config 2026-03-09T17:42:41.892 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:42:41 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-3[61096]: 2026-03-09T17:42:41.643+0000 7f823fe30740 -1 osd.3 0 log_to_monitors true 2026-03-09T17:42:41.893 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:41 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:41.893 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:41 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:41.893 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:41 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:41.893 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:41 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:41.893 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:41 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:41.893 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:41 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:41.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:41 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:41.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:41 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:41.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:41 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:41.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:41 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:41.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:41 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:41.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:41 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:42.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:41 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:42.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:41 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:42.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:41 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:42.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:41 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:42.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:41 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:42.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:41 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:42.702 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:42:42.724 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch daemon add osd vm04:/dev/vdc 2026-03-09T17:42:42.906 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.b/config 2026-03-09T17:42:42.930 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:42 vm04 ceph-mon[53158]: pgmap v47: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:42.930 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:42 vm04 ceph-mon[53158]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T17:42:42.930 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:42 vm04 ceph-mon[53158]: Detected new or changed devices on vm04 2026-03-09T17:42:42.930 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:42 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:42.930 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:42 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:42.930 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:42 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:42.930 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:42 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:42.930 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:42 vm04 ceph-mon[53158]: Adjusting osd_memory_target on vm04 to 2176M 2026-03-09T17:42:42.930 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:42 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:42.930 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:42 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:42.930 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:42 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:42.930 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:42 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:42.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:42 vm01 ceph-mon[52793]: pgmap v47: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:42.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:42 vm01 ceph-mon[52793]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T17:42:42.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:42 vm01 ceph-mon[52793]: Detected new or changed devices on vm04 2026-03-09T17:42:42.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:42 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:42.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:42 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:42.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:42 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:42.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:42 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:42.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:42 vm01 ceph-mon[52793]: Adjusting osd_memory_target on vm04 to 2176M 2026-03-09T17:42:42.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:42 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:42.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:42 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:42.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:42 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:42.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:42 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:43.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:42 vm05 ceph-mon[53831]: pgmap v47: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:43.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:42 vm05 ceph-mon[53831]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T17:42:43.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:42 vm05 ceph-mon[53831]: Detected new or changed devices on vm04 2026-03-09T17:42:43.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:42 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:43.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:42 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:43.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:42 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:43.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:42 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:43.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:42 vm05 ceph-mon[53831]: Adjusting osd_memory_target on vm04 to 2176M 2026-03-09T17:42:43.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:42 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:43.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:42 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:43.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:42 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:43.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:42 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:43.640 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:43 vm04 ceph-mon[53158]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-09T17:42:43.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:43 vm01 ceph-mon[52793]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-09T17:42:43.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:43 vm01 ceph-mon[52793]: osdmap e23: 4 total, 3 up, 4 in 2026-03-09T17:42:43.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:43 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:43.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:43 vm01 ceph-mon[52793]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T17:42:43.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:43 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:43.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:43 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:43.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:43 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:44.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:43 vm04 ceph-mon[53158]: osdmap e23: 4 total, 3 up, 4 in 2026-03-09T17:42:44.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:43 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:44.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:43 vm04 ceph-mon[53158]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T17:42:44.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:43 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:44.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:43 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:44.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:43 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:44.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:43 vm05 ceph-mon[53831]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-09T17:42:44.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:43 vm05 ceph-mon[53831]: osdmap e23: 4 total, 3 up, 4 in 2026-03-09T17:42:44.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:43 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:44.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:43 vm05 ceph-mon[53831]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T17:42:44.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:43 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:44.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:43 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:44.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:43 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:44.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: from='client.24206 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:44.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: pgmap v49: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:44.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T17:42:44.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: osdmap e24: 4 total, 3 up, 4 in 2026-03-09T17:42:44.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:44.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:44.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f2ea538a-0dea-48f3-87d2-b20de90a8311"}]: dispatch 2026-03-09T17:42:44.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.104:0/1819551419' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f2ea538a-0dea-48f3-87d2-b20de90a8311"}]: dispatch 2026-03-09T17:42:44.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "f2ea538a-0dea-48f3-87d2-b20de90a8311"}]': finished 2026-03-09T17:42:44.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: osdmap e25: 5 total, 3 up, 5 in 2026-03-09T17:42:44.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:44.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:44.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.104:0/4281485568' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:44.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' 2026-03-09T17:42:44.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:44 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:45.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: from='client.24206 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:45.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: pgmap v49: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:45.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T17:42:45.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: osdmap e24: 4 total, 3 up, 4 in 2026-03-09T17:42:45.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:45.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:45.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f2ea538a-0dea-48f3-87d2-b20de90a8311"}]: dispatch 2026-03-09T17:42:45.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.104:0/1819551419' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f2ea538a-0dea-48f3-87d2-b20de90a8311"}]: dispatch 2026-03-09T17:42:45.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "f2ea538a-0dea-48f3-87d2-b20de90a8311"}]': finished 2026-03-09T17:42:45.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: osdmap e25: 5 total, 3 up, 5 in 2026-03-09T17:42:45.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:45.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:45.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.104:0/4281485568' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:45.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' 2026-03-09T17:42:45.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:44 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:45.042 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:42:44 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-3[61096]: 2026-03-09T17:42:44.611+0000 7f823bdb1640 -1 osd.3 0 waiting for initial osdmap 2026-03-09T17:42:45.042 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:42:44 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-3[61096]: 2026-03-09T17:42:44.618+0000 7f82373da640 -1 osd.3 25 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:42:45.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: from='client.24206 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:45.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: pgmap v49: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:45.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T17:42:45.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: osdmap e24: 4 total, 3 up, 4 in 2026-03-09T17:42:45.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:45.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:45.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f2ea538a-0dea-48f3-87d2-b20de90a8311"}]: dispatch 2026-03-09T17:42:45.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.104:0/1819551419' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "f2ea538a-0dea-48f3-87d2-b20de90a8311"}]: dispatch 2026-03-09T17:42:45.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "f2ea538a-0dea-48f3-87d2-b20de90a8311"}]': finished 2026-03-09T17:42:45.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: osdmap e25: 5 total, 3 up, 5 in 2026-03-09T17:42:45.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:45.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:45.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.104:0/4281485568' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:45.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: from='osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626]' entity='osd.3' 2026-03-09T17:42:45.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:44 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:45 vm01 ceph-mon[52793]: purged_snaps scrub starts 2026-03-09T17:42:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:45 vm01 ceph-mon[52793]: purged_snaps scrub ok 2026-03-09T17:42:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:45 vm01 ceph-mon[52793]: osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626] boot 2026-03-09T17:42:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:45 vm01 ceph-mon[52793]: osdmap e26: 5 total, 4 up, 5 in 2026-03-09T17:42:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:45 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:45 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:46.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:45 vm04 ceph-mon[53158]: purged_snaps scrub starts 2026-03-09T17:42:46.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:45 vm04 ceph-mon[53158]: purged_snaps scrub ok 2026-03-09T17:42:46.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:45 vm04 ceph-mon[53158]: osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626] boot 2026-03-09T17:42:46.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:45 vm04 ceph-mon[53158]: osdmap e26: 5 total, 4 up, 5 in 2026-03-09T17:42:46.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:45 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:46.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:45 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:46.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:45 vm05 ceph-mon[53831]: purged_snaps scrub starts 2026-03-09T17:42:46.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:45 vm05 ceph-mon[53831]: purged_snaps scrub ok 2026-03-09T17:42:46.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:45 vm05 ceph-mon[53831]: osd.3 [v2:192.168.123.104:6808/2640539626,v1:192.168.123.104:6809/2640539626] boot 2026-03-09T17:42:46.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:45 vm05 ceph-mon[53831]: osdmap e26: 5 total, 4 up, 5 in 2026-03-09T17:42:46.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:45 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:42:46.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:45 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:47.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:47 vm04 ceph-mon[53158]: pgmap v53: 1 pgs: 1 remapped+peering; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:47.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:47 vm04 ceph-mon[53158]: osdmap e27: 5 total, 4 up, 5 in 2026-03-09T17:42:47.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:47 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:47.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:47 vm05 ceph-mon[53831]: pgmap v53: 1 pgs: 1 remapped+peering; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:47.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:47 vm05 ceph-mon[53831]: osdmap e27: 5 total, 4 up, 5 in 2026-03-09T17:42:47.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:47 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:47.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:47 vm01 ceph-mon[52793]: pgmap v53: 1 pgs: 1 remapped+peering; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:42:47.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:47 vm01 ceph-mon[52793]: osdmap e27: 5 total, 4 up, 5 in 2026-03-09T17:42:47.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:47 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:48.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:48 vm04 ceph-mon[53158]: osdmap e28: 5 total, 4 up, 5 in 2026-03-09T17:42:48.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:48 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:48.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:48 vm05 ceph-mon[53831]: osdmap e28: 5 total, 4 up, 5 in 2026-03-09T17:42:48.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:48 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:48.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:48 vm01 ceph-mon[52793]: osdmap e28: 5 total, 4 up, 5 in 2026-03-09T17:42:48.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:48 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:49.202 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:49 vm04 ceph-mon[53158]: pgmap v56: 1 pgs: 1 remapped+peering; 449 KiB data, 507 MiB used, 79 GiB / 80 GiB avail 2026-03-09T17:42:49.203 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:49 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-09T17:42:49.203 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:49 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:49.203 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:49 vm04 ceph-mon[53158]: Deploying daemon osd.4 on vm04 2026-03-09T17:42:49.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:49 vm05 ceph-mon[53831]: pgmap v56: 1 pgs: 1 remapped+peering; 449 KiB data, 507 MiB used, 79 GiB / 80 GiB avail 2026-03-09T17:42:49.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:49 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-09T17:42:49.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:49 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:49.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:49 vm05 ceph-mon[53831]: Deploying daemon osd.4 on vm04 2026-03-09T17:42:49.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:49 vm01 ceph-mon[52793]: pgmap v56: 1 pgs: 1 remapped+peering; 449 KiB data, 507 MiB used, 79 GiB / 80 GiB avail 2026-03-09T17:42:49.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:49 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-09T17:42:49.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:49 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:49.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:49 vm01 ceph-mon[52793]: Deploying daemon osd.4 on vm04 2026-03-09T17:42:51.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:51 vm05 ceph-mon[53831]: pgmap v57: 1 pgs: 1 remapped+peering; 449 KiB data, 507 MiB used, 79 GiB / 80 GiB avail 2026-03-09T17:42:51.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:51 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:51.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:51 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:51.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:51 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:51 vm01 ceph-mon[52793]: pgmap v57: 1 pgs: 1 remapped+peering; 449 KiB data, 507 MiB used, 79 GiB / 80 GiB avail 2026-03-09T17:42:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:51 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:51 vm04 ceph-mon[53158]: pgmap v57: 1 pgs: 1 remapped+peering; 449 KiB data, 507 MiB used, 79 GiB / 80 GiB avail 2026-03-09T17:42:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:51 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:51 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:51 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:51.635 INFO:teuthology.orchestra.run.vm04.stdout:Created osd(s) 4 on host 'vm04' 2026-03-09T17:42:51.692 DEBUG:teuthology.orchestra.run.vm04:osd.4> sudo journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.4.service 2026-03-09T17:42:51.694 INFO:tasks.cephadm:Deploying osd.5 on vm05 with /dev/vde... 2026-03-09T17:42:51.694 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- lvm zap /dev/vde 2026-03-09T17:42:51.869 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.c/config 2026-03-09T17:42:52.147 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:42:52 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-4[66342]: 2026-03-09T17:42:52.092+0000 7f5bd972d740 -1 osd.4 0 log_to_monitors true 2026-03-09T17:42:52.304 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:52 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.304 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:52 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.304 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:52 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:52.304 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:52 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:52.304 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:52 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.304 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:52 vm05 ceph-mon[53831]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 507 MiB used, 79 GiB / 80 GiB avail; 70 KiB/s, 0 objects/s recovering 2026-03-09T17:42:52.304 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:52 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:52.304 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:52 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.304 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:52 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.304 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:52 vm05 ceph-mon[53831]: from='osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T17:42:52.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:52 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:52 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:52 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:52.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:52 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:52.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:52 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:52 vm01 ceph-mon[52793]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 507 MiB used, 79 GiB / 80 GiB avail; 70 KiB/s, 0 objects/s recovering 2026-03-09T17:42:52.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:52 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:52.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:52 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:52 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:52 vm01 ceph-mon[52793]: from='osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T17:42:52.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:52 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:52 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:52 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:52.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:52 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:52.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:52 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.542 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:52 vm04 ceph-mon[53158]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 507 MiB used, 79 GiB / 80 GiB avail; 70 KiB/s, 0 objects/s recovering 2026-03-09T17:42:52.542 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:52 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:52.542 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:52 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.542 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:52 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:52.542 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:52 vm04 ceph-mon[53158]: from='osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T17:42:52.736 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:42:52.763 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch daemon add osd vm05:/dev/vde 2026-03-09T17:42:52.952 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.c/config 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262]' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: osdmap e29: 5 total, 4 up, 5 in 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: Detected new or changed devices on vm04 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: Adjusting osd_memory_target on vm04 to 1451M 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:53.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:53 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:53.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262]' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-09T17:42:53.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: osdmap e29: 5 total, 4 up, 5 in 2026-03-09T17:42:53.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T17:42:53.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:53.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: Detected new or changed devices on vm04 2026-03-09T17:42:53.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:53.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:53.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:53.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:53.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:53.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: Adjusting osd_memory_target on vm04 to 1451M 2026-03-09T17:42:53.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:53.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:53.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:53.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:53.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:53.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:53.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:53 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262]' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: osdmap e29: 5 total, 4 up, 5 in 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: Detected new or changed devices on vm04 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: Adjusting osd_memory_target on vm04 to 1451M 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:42:54.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:53 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:54.041 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:42:53 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-4[66342]: 2026-03-09T17:42:53.656+0000 7f5bd56ae640 -1 osd.4 0 waiting for initial osdmap 2026-03-09T17:42:54.041 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:42:53 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-4[66342]: 2026-03-09T17:42:53.662+0000 7f5bd14d8640 -1 osd.4 30 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:42:54.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:54 vm05 ceph-mon[53831]: from='client.24214 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:54.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:54 vm05 ceph-mon[53831]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 507 MiB used, 79 GiB / 80 GiB avail; 61 KiB/s, 0 objects/s recovering 2026-03-09T17:42:54.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:54 vm05 ceph-mon[53831]: from='osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262]' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T17:42:54.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:54 vm05 ceph-mon[53831]: osdmap e30: 5 total, 4 up, 5 in 2026-03-09T17:42:54.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:54 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:54.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:54 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:54.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:54 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/2173242411' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4dcf9f68-3a19-43f6-b608-2d0481914d8a"}]: dispatch 2026-03-09T17:42:54.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:54 vm05 ceph-mon[53831]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4dcf9f68-3a19-43f6-b608-2d0481914d8a"}]: dispatch 2026-03-09T17:42:54.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:54 vm05 ceph-mon[53831]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4dcf9f68-3a19-43f6-b608-2d0481914d8a"}]': finished 2026-03-09T17:42:54.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:54 vm05 ceph-mon[53831]: osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262] boot 2026-03-09T17:42:54.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:54 vm05 ceph-mon[53831]: osdmap e31: 6 total, 5 up, 6 in 2026-03-09T17:42:54.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:54 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:54.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:54 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:42:54.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:54 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/610829876' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:54.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:54 vm01 ceph-mon[52793]: from='client.24214 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:54.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:54 vm01 ceph-mon[52793]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 507 MiB used, 79 GiB / 80 GiB avail; 61 KiB/s, 0 objects/s recovering 2026-03-09T17:42:54.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:54 vm01 ceph-mon[52793]: from='osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262]' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T17:42:54.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:54 vm01 ceph-mon[52793]: osdmap e30: 5 total, 4 up, 5 in 2026-03-09T17:42:54.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:54 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:54.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:54 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:54.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:54 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/2173242411' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4dcf9f68-3a19-43f6-b608-2d0481914d8a"}]: dispatch 2026-03-09T17:42:54.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:54 vm01 ceph-mon[52793]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4dcf9f68-3a19-43f6-b608-2d0481914d8a"}]: dispatch 2026-03-09T17:42:54.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:54 vm01 ceph-mon[52793]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4dcf9f68-3a19-43f6-b608-2d0481914d8a"}]': finished 2026-03-09T17:42:54.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:54 vm01 ceph-mon[52793]: osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262] boot 2026-03-09T17:42:54.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:54 vm01 ceph-mon[52793]: osdmap e31: 6 total, 5 up, 6 in 2026-03-09T17:42:54.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:54 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:54.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:54 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:42:54.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:54 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/610829876' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:54 vm04 ceph-mon[53158]: from='client.24214 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:42:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:54 vm04 ceph-mon[53158]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 507 MiB used, 79 GiB / 80 GiB avail; 61 KiB/s, 0 objects/s recovering 2026-03-09T17:42:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:54 vm04 ceph-mon[53158]: from='osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262]' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-09T17:42:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:54 vm04 ceph-mon[53158]: osdmap e30: 5 total, 4 up, 5 in 2026-03-09T17:42:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:54 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:54 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:54 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/2173242411' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4dcf9f68-3a19-43f6-b608-2d0481914d8a"}]: dispatch 2026-03-09T17:42:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:54 vm04 ceph-mon[53158]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4dcf9f68-3a19-43f6-b608-2d0481914d8a"}]: dispatch 2026-03-09T17:42:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:54 vm04 ceph-mon[53158]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4dcf9f68-3a19-43f6-b608-2d0481914d8a"}]': finished 2026-03-09T17:42:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:54 vm04 ceph-mon[53158]: osd.4 [v2:192.168.123.104:6816/3638700262,v1:192.168.123.104:6817/3638700262] boot 2026-03-09T17:42:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:54 vm04 ceph-mon[53158]: osdmap e31: 6 total, 5 up, 6 in 2026-03-09T17:42:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:54 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:42:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:54 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:42:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:54 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/610829876' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:42:56.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:56 vm05 ceph-mon[53831]: purged_snaps scrub starts 2026-03-09T17:42:56.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:56 vm05 ceph-mon[53831]: purged_snaps scrub ok 2026-03-09T17:42:56.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:56 vm05 ceph-mon[53831]: osdmap e32: 6 total, 5 up, 6 in 2026-03-09T17:42:56.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:56 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:42:56.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:56 vm01 ceph-mon[52793]: purged_snaps scrub starts 2026-03-09T17:42:56.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:56 vm01 ceph-mon[52793]: purged_snaps scrub ok 2026-03-09T17:42:56.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:56 vm01 ceph-mon[52793]: osdmap e32: 6 total, 5 up, 6 in 2026-03-09T17:42:56.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:56 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:42:56.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:56 vm04 ceph-mon[53158]: purged_snaps scrub starts 2026-03-09T17:42:56.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:56 vm04 ceph-mon[53158]: purged_snaps scrub ok 2026-03-09T17:42:56.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:56 vm04 ceph-mon[53158]: osdmap e32: 6 total, 5 up, 6 in 2026-03-09T17:42:56.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:56 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:42:57.260 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:57 vm05 ceph-mon[53831]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail 2026-03-09T17:42:57.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:57 vm01 ceph-mon[52793]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail 2026-03-09T17:42:57.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:57 vm04 ceph-mon[53158]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail 2026-03-09T17:42:58.207 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:58 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-09T17:42:58.207 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:58 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:58.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:58 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-09T17:42:58.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:58 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:58.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:58 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-09T17:42:58.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:58 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:59.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:59 vm05 ceph-mon[53831]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:42:59.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:42:59 vm05 ceph-mon[53831]: Deploying daemon osd.5 on vm05 2026-03-09T17:42:59.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:59 vm01 ceph-mon[52793]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:42:59.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:42:59 vm01 ceph-mon[52793]: Deploying daemon osd.5 on vm05 2026-03-09T17:42:59.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:59 vm04 ceph-mon[53158]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:42:59.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:42:59 vm04 ceph-mon[53158]: Deploying daemon osd.5 on vm05 2026-03-09T17:43:01.313 INFO:teuthology.orchestra.run.vm05.stdout:Created osd(s) 5 on host 'vm05' 2026-03-09T17:43:01.326 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:01 vm05 ceph-mon[53831]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:43:01.326 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:01 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:01.326 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:01 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.326 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:01 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.326 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:01 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.326 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:01 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.326 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:01 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:01.326 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:01 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:01.326 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:01 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.378 DEBUG:teuthology.orchestra.run.vm05:osd.5> sudo journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.5.service 2026-03-09T17:43:01.379 INFO:tasks.cephadm:Deploying osd.6 on vm05 with /dev/vdd... 2026-03-09T17:43:01.379 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- lvm zap /dev/vdd 2026-03-09T17:43:01.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:01 vm01 ceph-mon[52793]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:43:01.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:01 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:01.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:01 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:01 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:01 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:01 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:01 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:01.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:01 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:01.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:01 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:01 vm04 ceph-mon[53158]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:43:01.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:01 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:01.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:01 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:01 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:01 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:01 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:01 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:01.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:01 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:01.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:01 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:01.586 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:43:01 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-5[57405]: 2026-03-09T17:43:01.492+0000 7f706f1de740 -1 Falling back to public interface 2026-03-09T17:43:01.682 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.c/config 2026-03-09T17:43:02.329 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:02 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:02.329 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:02 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:02.329 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:02 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:02.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:02 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:02.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:02 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:02.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:02 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:02.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:02 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:02.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:02 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:02.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:02 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:02.958 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:43:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-5[57405]: 2026-03-09T17:43:02.578+0000 7f706f1de740 -1 osd.5 0 log_to_monitors true 2026-03-09T17:43:03.068 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:43:03.096 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch daemon add osd vm05:/dev/vdd 2026-03-09T17:43:03.274 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.c/config 2026-03-09T17:43:03.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:03 vm05 ceph-mon[53831]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:43:03.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:03 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:03.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:03 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:03.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:03 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:03.303 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:03 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:03.303 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:03 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:03.303 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:03 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:03.303 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:03 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:03.303 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:03 vm05 ceph-mon[53831]: from='osd.5 [v2:192.168.123.105:6800/2510326788,v1:192.168.123.105:6801/2510326788]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T17:43:03.303 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:03 vm05 ceph-mon[53831]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T17:43:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:03 vm01 ceph-mon[52793]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:43:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:03 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:03 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:03 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:03 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:03 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:03 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:03 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:03 vm01 ceph-mon[52793]: from='osd.5 [v2:192.168.123.105:6800/2510326788,v1:192.168.123.105:6801/2510326788]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T17:43:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:03 vm01 ceph-mon[52793]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T17:43:03.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:03 vm04 ceph-mon[53158]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:43:03.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:03 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:03.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:03 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:03.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:03 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:03.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:03 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:03.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:03 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:03.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:03 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:03.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:03 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:03.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:03 vm04 ceph-mon[53158]: from='osd.5 [v2:192.168.123.105:6800/2510326788,v1:192.168.123.105:6801/2510326788]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T17:43:03.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:03 vm04 ceph-mon[53158]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T17:43:04.334 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:04 vm05 ceph-mon[53831]: Detected new or changed devices on vm05 2026-03-09T17:43:04.335 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:04 vm05 ceph-mon[53831]: Adjusting osd_memory_target on vm05 to 4353M 2026-03-09T17:43:04.335 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:04 vm05 ceph-mon[53831]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-09T17:43:04.335 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:04 vm05 ceph-mon[53831]: osdmap e33: 6 total, 5 up, 6 in 2026-03-09T17:43:04.335 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:04 vm05 ceph-mon[53831]: from='osd.5 [v2:192.168.123.105:6800/2510326788,v1:192.168.123.105:6801/2510326788]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:04.335 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:04 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:43:04.335 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:04 vm05 ceph-mon[53831]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:04.335 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:04 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:43:04.335 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:04 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:43:04.335 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:04 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:04.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:04 vm01 ceph-mon[52793]: Detected new or changed devices on vm05 2026-03-09T17:43:04.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:04 vm01 ceph-mon[52793]: Adjusting osd_memory_target on vm05 to 4353M 2026-03-09T17:43:04.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:04 vm01 ceph-mon[52793]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-09T17:43:04.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:04 vm01 ceph-mon[52793]: osdmap e33: 6 total, 5 up, 6 in 2026-03-09T17:43:04.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:04 vm01 ceph-mon[52793]: from='osd.5 [v2:192.168.123.105:6800/2510326788,v1:192.168.123.105:6801/2510326788]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:04.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:04 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:43:04.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:04 vm01 ceph-mon[52793]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:04.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:04 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:43:04.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:04 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:43:04.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:04 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:04 vm04 ceph-mon[53158]: Detected new or changed devices on vm05 2026-03-09T17:43:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:04 vm04 ceph-mon[53158]: Adjusting osd_memory_target on vm05 to 4353M 2026-03-09T17:43:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:04 vm04 ceph-mon[53158]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-09T17:43:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:04 vm04 ceph-mon[53158]: osdmap e33: 6 total, 5 up, 6 in 2026-03-09T17:43:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:04 vm04 ceph-mon[53158]: from='osd.5 [v2:192.168.123.105:6800/2510326788,v1:192.168.123.105:6801/2510326788]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:04 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:43:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:04 vm04 ceph-mon[53158]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:04 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:43:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:04 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:43:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:04 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:04.598 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:43:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-5[57405]: 2026-03-09T17:43:04.331+0000 7f706b972640 -1 osd.5 0 waiting for initial osdmap 2026-03-09T17:43:04.598 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:43:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-5[57405]: 2026-03-09T17:43:04.338+0000 7f7066f89640 -1 osd.5 34 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:43:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:05 vm01 ceph-mon[52793]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:43:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:05 vm01 ceph-mon[52793]: from='client.24248 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:05 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1927343141' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "136fb72c-868c-451b-a36d-2e332f552fff"}]: dispatch 2026-03-09T17:43:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:05 vm01 ceph-mon[52793]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-09T17:43:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:05 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1927343141' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "136fb72c-868c-451b-a36d-2e332f552fff"}]': finished 2026-03-09T17:43:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:05 vm01 ceph-mon[52793]: osdmap e34: 7 total, 5 up, 7 in 2026-03-09T17:43:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:05 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:43:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:05 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:05 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:43:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:05 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1596099793' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:43:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:05 vm01 ceph-mon[52793]: osd.5 [v2:192.168.123.105:6800/2510326788,v1:192.168.123.105:6801/2510326788] boot 2026-03-09T17:43:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:05 vm01 ceph-mon[52793]: osdmap e35: 7 total, 6 up, 7 in 2026-03-09T17:43:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:05 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:43:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:05 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:05 vm04 ceph-mon[53158]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:43:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:05 vm04 ceph-mon[53158]: from='client.24248 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:05 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1927343141' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "136fb72c-868c-451b-a36d-2e332f552fff"}]: dispatch 2026-03-09T17:43:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:05 vm04 ceph-mon[53158]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-09T17:43:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:05 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1927343141' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "136fb72c-868c-451b-a36d-2e332f552fff"}]': finished 2026-03-09T17:43:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:05 vm04 ceph-mon[53158]: osdmap e34: 7 total, 5 up, 7 in 2026-03-09T17:43:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:05 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:43:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:05 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:05 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:43:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:05 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1596099793' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:43:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:05 vm04 ceph-mon[53158]: osd.5 [v2:192.168.123.105:6800/2510326788,v1:192.168.123.105:6801/2510326788] boot 2026-03-09T17:43:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:05 vm04 ceph-mon[53158]: osdmap e35: 7 total, 6 up, 7 in 2026-03-09T17:43:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:05 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:43:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:05 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:05 vm05 ceph-mon[53831]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:43:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:05 vm05 ceph-mon[53831]: from='client.24248 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:05 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1927343141' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "136fb72c-868c-451b-a36d-2e332f552fff"}]: dispatch 2026-03-09T17:43:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:05 vm05 ceph-mon[53831]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-09T17:43:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:05 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1927343141' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "136fb72c-868c-451b-a36d-2e332f552fff"}]': finished 2026-03-09T17:43:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:05 vm05 ceph-mon[53831]: osdmap e34: 7 total, 5 up, 7 in 2026-03-09T17:43:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:05 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:43:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:05 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:05 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:43:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:05 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1596099793' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:43:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:05 vm05 ceph-mon[53831]: osd.5 [v2:192.168.123.105:6800/2510326788,v1:192.168.123.105:6801/2510326788] boot 2026-03-09T17:43:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:05 vm05 ceph-mon[53831]: osdmap e35: 7 total, 6 up, 7 in 2026-03-09T17:43:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:05 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:43:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:05 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:06.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:06 vm01 ceph-mon[52793]: purged_snaps scrub starts 2026-03-09T17:43:06.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:06 vm01 ceph-mon[52793]: purged_snaps scrub ok 2026-03-09T17:43:06.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:06 vm01 ceph-mon[52793]: osdmap e36: 7 total, 6 up, 7 in 2026-03-09T17:43:06.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:06 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:06.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:06 vm04 ceph-mon[53158]: purged_snaps scrub starts 2026-03-09T17:43:06.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:06 vm04 ceph-mon[53158]: purged_snaps scrub ok 2026-03-09T17:43:06.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:06 vm04 ceph-mon[53158]: osdmap e36: 7 total, 6 up, 7 in 2026-03-09T17:43:06.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:06 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:06 vm05 ceph-mon[53831]: purged_snaps scrub starts 2026-03-09T17:43:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:06 vm05 ceph-mon[53831]: purged_snaps scrub ok 2026-03-09T17:43:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:06 vm05 ceph-mon[53831]: osdmap e36: 7 total, 6 up, 7 in 2026-03-09T17:43:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:06 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:07.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:07 vm01 ceph-mon[52793]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:43:07.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:07 vm01 ceph-mon[52793]: osdmap e37: 7 total, 6 up, 7 in 2026-03-09T17:43:07.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:07 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:07.482 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:07 vm05 ceph-mon[53831]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:43:07.482 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:07 vm05 ceph-mon[53831]: osdmap e37: 7 total, 6 up, 7 in 2026-03-09T17:43:07.482 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:07 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:07.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:07 vm04 ceph-mon[53158]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:43:07.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:07 vm04 ceph-mon[53158]: osdmap e37: 7 total, 6 up, 7 in 2026-03-09T17:43:07.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:07 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:09.270 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:09 vm05 ceph-mon[53831]: pgmap v75: 1 pgs: 1 remapped+peering; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:43:09.271 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:09 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-09T17:43:09.271 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:09 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:09.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:09 vm01 ceph-mon[52793]: pgmap v75: 1 pgs: 1 remapped+peering; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:43:09.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:09 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-09T17:43:09.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:09 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:09.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:09 vm04 ceph-mon[53158]: pgmap v75: 1 pgs: 1 remapped+peering; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:43:09.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:09 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-09T17:43:09.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:09 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:10.451 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:10 vm05 ceph-mon[53831]: Deploying daemon osd.6 on vm05 2026-03-09T17:43:10.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:10 vm01 ceph-mon[52793]: Deploying daemon osd.6 on vm05 2026-03-09T17:43:10.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:10 vm04 ceph-mon[53158]: Deploying daemon osd.6 on vm05 2026-03-09T17:43:11.168 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:11 vm05 ceph-mon[53831]: pgmap v76: 1 pgs: 1 remapped+peering; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:43:11.168 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:11 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:11.168 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:11 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:11.168 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:11 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:11.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:11 vm01 ceph-mon[52793]: pgmap v76: 1 pgs: 1 remapped+peering; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:43:11.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:11 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:11.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:11 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:11.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:11 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:11.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:11 vm04 ceph-mon[53158]: pgmap v76: 1 pgs: 1 remapped+peering; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:43:11.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:11 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:11.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:11 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:11.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:11 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:11.701 INFO:teuthology.orchestra.run.vm05.stdout:Created osd(s) 6 on host 'vm05' 2026-03-09T17:43:11.763 DEBUG:teuthology.orchestra.run.vm05:osd.6> sudo journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.6.service 2026-03-09T17:43:11.764 INFO:tasks.cephadm:Deploying osd.7 on vm05 with /dev/vdc... 2026-03-09T17:43:11.764 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- lvm zap /dev/vdc 2026-03-09T17:43:12.060 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.c/config 2026-03-09T17:43:12.270 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:12 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:12.270 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:12 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:12.270 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:12 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:12.270 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:12 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:12.270 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:12 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:12.270 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:12 vm05 ceph-mon[53831]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail; 68 KiB/s, 0 objects/s recovering 2026-03-09T17:43:12.270 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:12 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:12.270 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:12 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:12.270 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:12 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:12.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:12 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:12.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:12 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:12.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:12 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:12.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:12 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:12.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:12 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:12.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:12 vm04 ceph-mon[53158]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail; 68 KiB/s, 0 objects/s recovering 2026-03-09T17:43:12.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:12 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:12.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:12 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:12.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:12 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:12.570 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:43:12 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-6[62565]: 2026-03-09T17:43:12.320+0000 7f6296dc2740 -1 osd.6 0 log_to_monitors true 2026-03-09T17:43:12.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:12 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:12.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:12 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:12.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:12 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:12.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:12 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:12.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:12 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:12.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:12 vm01 ceph-mon[52793]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail; 68 KiB/s, 0 objects/s recovering 2026-03-09T17:43:12.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:12 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:12.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:12 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:12.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:12 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:13.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:13 vm05 ceph-mon[53831]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-09T17:43:13.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:13 vm05 ceph-mon[53831]: from='osd.6 [v2:192.168.123.105:6808/4036098223,v1:192.168.123.105:6809/4036098223]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-09T17:43:13.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:13 vm05 ceph-mon[53831]: Detected new or changed devices on vm05 2026-03-09T17:43:13.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:13 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:13.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:13 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:13.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:13 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:13.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:13 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:13.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:13 vm05 ceph-mon[53831]: Adjusting osd_memory_target on vm05 to 2176M 2026-03-09T17:43:13.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:13 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:13.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:13 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:13.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:13 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:13.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:13 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:13.510 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:43:13.535 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch daemon add osd vm05:/dev/vdc 2026-03-09T17:43:13.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:13 vm04 ceph-mon[53158]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-09T17:43:13.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:13 vm04 ceph-mon[53158]: from='osd.6 [v2:192.168.123.105:6808/4036098223,v1:192.168.123.105:6809/4036098223]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-09T17:43:13.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:13 vm04 ceph-mon[53158]: Detected new or changed devices on vm05 2026-03-09T17:43:13.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:13 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:13.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:13 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:13.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:13 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:13.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:13 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:13.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:13 vm04 ceph-mon[53158]: Adjusting osd_memory_target on vm05 to 2176M 2026-03-09T17:43:13.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:13 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:13.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:13 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:13.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:13 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:13.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:13 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:13 vm01 ceph-mon[52793]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-09T17:43:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:13 vm01 ceph-mon[52793]: from='osd.6 [v2:192.168.123.105:6808/4036098223,v1:192.168.123.105:6809/4036098223]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-09T17:43:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:13 vm01 ceph-mon[52793]: Detected new or changed devices on vm05 2026-03-09T17:43:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:13 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:13 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:13 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:13 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:13 vm01 ceph-mon[52793]: Adjusting osd_memory_target on vm05 to 2176M 2026-03-09T17:43:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:13 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:13 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:13 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:13.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:13 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:13.735 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.c/config 2026-03-09T17:43:14.523 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:14 vm05 ceph-mon[53831]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-09T17:43:14.524 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:14 vm05 ceph-mon[53831]: osdmap e38: 7 total, 6 up, 7 in 2026-03-09T17:43:14.524 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:14 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:14.524 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:14 vm05 ceph-mon[53831]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:14.524 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:14 vm05 ceph-mon[53831]: from='osd.6 [v2:192.168.123.105:6808/4036098223,v1:192.168.123.105:6809/4036098223]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:14.524 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:14 vm05 ceph-mon[53831]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail; 59 KiB/s, 0 objects/s recovering 2026-03-09T17:43:14.524 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:14 vm05 ceph-mon[53831]: from='client.24268 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:14.524 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:14 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:43:14.524 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:14 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:43:14.524 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:14 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:14.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:14 vm01 ceph-mon[52793]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-09T17:43:14.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:14 vm01 ceph-mon[52793]: osdmap e38: 7 total, 6 up, 7 in 2026-03-09T17:43:14.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:14 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:14.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:14 vm01 ceph-mon[52793]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:14.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:14 vm01 ceph-mon[52793]: from='osd.6 [v2:192.168.123.105:6808/4036098223,v1:192.168.123.105:6809/4036098223]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:14.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:14 vm01 ceph-mon[52793]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail; 59 KiB/s, 0 objects/s recovering 2026-03-09T17:43:14.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:14 vm01 ceph-mon[52793]: from='client.24268 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:14.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:14 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:43:14.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:14 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:43:14.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:14 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:14 vm04 ceph-mon[53158]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-09T17:43:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:14 vm04 ceph-mon[53158]: osdmap e38: 7 total, 6 up, 7 in 2026-03-09T17:43:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:14 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:14 vm04 ceph-mon[53158]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:14 vm04 ceph-mon[53158]: from='osd.6 [v2:192.168.123.105:6808/4036098223,v1:192.168.123.105:6809/4036098223]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:14 vm04 ceph-mon[53158]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail; 59 KiB/s, 0 objects/s recovering 2026-03-09T17:43:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:14 vm04 ceph-mon[53158]: from='client.24268 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:14 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:43:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:14 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:43:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:14 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:15.544 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:15 vm05 ceph-mon[53831]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-09T17:43:15.545 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:15 vm05 ceph-mon[53831]: osdmap e39: 7 total, 6 up, 7 in 2026-03-09T17:43:15.545 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:15 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:15.545 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:15 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:15.545 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:15 vm05 ceph-mon[53831]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "11afd218-abae-4bf5-ac97-bba7cd66648d"}]: dispatch 2026-03-09T17:43:15.545 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:15 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3560640914' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "11afd218-abae-4bf5-ac97-bba7cd66648d"}]: dispatch 2026-03-09T17:43:15.545 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:15 vm05 ceph-mon[53831]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "11afd218-abae-4bf5-ac97-bba7cd66648d"}]': finished 2026-03-09T17:43:15.545 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:15 vm05 ceph-mon[53831]: osdmap e40: 8 total, 6 up, 8 in 2026-03-09T17:43:15.545 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:15 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:15.545 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:15 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:15.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:15 vm01 ceph-mon[52793]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-09T17:43:15.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:15 vm01 ceph-mon[52793]: osdmap e39: 7 total, 6 up, 7 in 2026-03-09T17:43:15.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:15 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:15.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:15 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:15.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:15 vm01 ceph-mon[52793]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "11afd218-abae-4bf5-ac97-bba7cd66648d"}]: dispatch 2026-03-09T17:43:15.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:15 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3560640914' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "11afd218-abae-4bf5-ac97-bba7cd66648d"}]: dispatch 2026-03-09T17:43:15.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:15 vm01 ceph-mon[52793]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "11afd218-abae-4bf5-ac97-bba7cd66648d"}]': finished 2026-03-09T17:43:15.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:15 vm01 ceph-mon[52793]: osdmap e40: 8 total, 6 up, 8 in 2026-03-09T17:43:15.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:15 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:15.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:15 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:15.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:15 vm04 ceph-mon[53158]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-09T17:43:15.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:15 vm04 ceph-mon[53158]: osdmap e39: 7 total, 6 up, 7 in 2026-03-09T17:43:15.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:15 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:15.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:15 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:15.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:15 vm04 ceph-mon[53158]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "11afd218-abae-4bf5-ac97-bba7cd66648d"}]: dispatch 2026-03-09T17:43:15.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:15 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3560640914' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "11afd218-abae-4bf5-ac97-bba7cd66648d"}]: dispatch 2026-03-09T17:43:15.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:15 vm04 ceph-mon[53158]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "11afd218-abae-4bf5-ac97-bba7cd66648d"}]': finished 2026-03-09T17:43:15.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:15 vm04 ceph-mon[53158]: osdmap e40: 8 total, 6 up, 8 in 2026-03-09T17:43:15.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:15 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:15.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:15 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:16.314 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:43:15 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-6[62565]: 2026-03-09T17:43:15.879+0000 7f6292d43640 -1 osd.6 0 waiting for initial osdmap 2026-03-09T17:43:16.314 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:43:15 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-6[62565]: 2026-03-09T17:43:15.890+0000 7f628e36c640 -1 osd.6 40 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:43:16.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:16 vm05 ceph-mon[53831]: purged_snaps scrub starts 2026-03-09T17:43:16.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:16 vm05 ceph-mon[53831]: purged_snaps scrub ok 2026-03-09T17:43:16.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:16 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:16.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:16 vm05 ceph-mon[53831]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-09T17:43:16.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:16 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/2292268795' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:43:16.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:16 vm05 ceph-mon[53831]: from='osd.6 ' entity='osd.6' 2026-03-09T17:43:16.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:16 vm01 ceph-mon[52793]: purged_snaps scrub starts 2026-03-09T17:43:16.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:16 vm01 ceph-mon[52793]: purged_snaps scrub ok 2026-03-09T17:43:16.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:16 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:16.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:16 vm01 ceph-mon[52793]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-09T17:43:16.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:16 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/2292268795' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:43:16.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:16 vm01 ceph-mon[52793]: from='osd.6 ' entity='osd.6' 2026-03-09T17:43:16.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:16 vm04 ceph-mon[53158]: purged_snaps scrub starts 2026-03-09T17:43:16.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:16 vm04 ceph-mon[53158]: purged_snaps scrub ok 2026-03-09T17:43:16.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:16 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:16.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:16 vm04 ceph-mon[53158]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-09T17:43:16.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:16 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/2292268795' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:43:16.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:16 vm04 ceph-mon[53158]: from='osd.6 ' entity='osd.6' 2026-03-09T17:43:17.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:17 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:17.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:17 vm05 ceph-mon[53831]: osd.6 [v2:192.168.123.105:6808/4036098223,v1:192.168.123.105:6809/4036098223] boot 2026-03-09T17:43:17.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:17 vm05 ceph-mon[53831]: osdmap e41: 8 total, 7 up, 8 in 2026-03-09T17:43:17.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:17 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:17.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:17 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:17.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:17 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:17.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:17 vm01 ceph-mon[52793]: osd.6 [v2:192.168.123.105:6808/4036098223,v1:192.168.123.105:6809/4036098223] boot 2026-03-09T17:43:17.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:17 vm01 ceph-mon[52793]: osdmap e41: 8 total, 7 up, 8 in 2026-03-09T17:43:17.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:17 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:17.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:17 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:17.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:17 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:17.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:17 vm04 ceph-mon[53158]: osd.6 [v2:192.168.123.105:6808/4036098223,v1:192.168.123.105:6809/4036098223] boot 2026-03-09T17:43:17.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:17 vm04 ceph-mon[53158]: osdmap e41: 8 total, 7 up, 8 in 2026-03-09T17:43:17.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:17 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:43:17.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:17 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:18.582 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:18 vm05 ceph-mon[53831]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:18.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:18 vm01 ceph-mon[52793]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:18.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:18 vm04 ceph-mon[53158]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:19.497 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:19 vm05 ceph-mon[53831]: osdmap e42: 8 total, 7 up, 8 in 2026-03-09T17:43:19.497 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:19 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:19.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:19 vm01 ceph-mon[52793]: osdmap e42: 8 total, 7 up, 8 in 2026-03-09T17:43:19.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:19 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:19.790 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:19 vm04 ceph-mon[53158]: osdmap e42: 8 total, 7 up, 8 in 2026-03-09T17:43:19.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:19 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:20.619 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:20 vm05 ceph-mon[53831]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:20.619 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:20 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-09T17:43:20.619 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:20 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:20.619 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:20 vm05 ceph-mon[53831]: Deploying daemon osd.7 on vm05 2026-03-09T17:43:20.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:20 vm01 ceph-mon[52793]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:20.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:20 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-09T17:43:20.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:20 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:20.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:20 vm01 ceph-mon[52793]: Deploying daemon osd.7 on vm05 2026-03-09T17:43:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:20 vm04 ceph-mon[53158]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:20 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-09T17:43:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:20 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:20 vm04 ceph-mon[53158]: Deploying daemon osd.7 on vm05 2026-03-09T17:43:22.568 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:22 vm05 ceph-mon[53831]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:22.568 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:22 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:22.568 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:22 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:22.568 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:22 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:22.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:22 vm01 ceph-mon[52793]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:22.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:22 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:22.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:22 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:22.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:22 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:22.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:22 vm04 ceph-mon[53158]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:22.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:22 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:22.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:22 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:22.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:22 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:22.874 INFO:teuthology.orchestra.run.vm05.stdout:Created osd(s) 7 on host 'vm05' 2026-03-09T17:43:22.939 DEBUG:teuthology.orchestra.run.vm05:osd.7> sudo journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.7.service 2026-03-09T17:43:22.941 INFO:tasks.cephadm:Waiting for 8 OSDs to come up... 2026-03-09T17:43:22.941 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd stat -f json 2026-03-09T17:43:23.109 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:23.348 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:43:23.405 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:43:23 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-7[67389]: 2026-03-09T17:43:23.402+0000 7f752a0be740 -1 osd.7 0 log_to_monitors true 2026-03-09T17:43:23.408 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":42,"num_osds":8,"num_up_osds":7,"osd_up_since":1773078196,"num_in_osds":8,"osd_in_since":1773078195,"num_remapped_pgs":0} 2026-03-09T17:43:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:23 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:23 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:23 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:23 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:23 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:23 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:23 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:23 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:23 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3264359192' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:43:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:23 vm01 ceph-mon[52793]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-09T17:43:23.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:23 vm01 ceph-mon[52793]: from='osd.7 [v2:192.168.123.105:6816/2022540299,v1:192.168.123.105:6817/2022540299]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-09T17:43:23.758 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:23 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.758 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:23 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.758 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:23 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:23.758 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:23 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:23.758 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:23 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.758 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:23 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:23.758 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:23 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.758 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:23 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.758 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:23 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3264359192' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:43:23.758 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:23 vm05 ceph-mon[53831]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-09T17:43:23.758 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:23 vm05 ceph-mon[53831]: from='osd.7 [v2:192.168.123.105:6816/2022540299,v1:192.168.123.105:6817/2022540299]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-09T17:43:23.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:23 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:23 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:23 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:23.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:23 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:23.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:23 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:23 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:23.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:23 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:23 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:23.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:23 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3264359192' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:43:23.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:23 vm04 ceph-mon[53158]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-09T17:43:23.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:23 vm04 ceph-mon[53158]: from='osd.7 [v2:192.168.123.105:6816/2022540299,v1:192.168.123.105:6817/2022540299]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-09T17:43:24.408 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd stat -f json 2026-03-09T17:43:24.589 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:24.815 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:43:24.890 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":43,"num_osds":8,"num_up_osds":7,"osd_up_since":1773078196,"num_in_osds":8,"osd_in_since":1773078195,"num_remapped_pgs":0} 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: Detected new or changed devices on vm05 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: Adjusting osd_memory_target on vm05 to 1451M 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: osdmap e43: 8 total, 7 up, 8 in 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: from='osd.7 [v2:192.168.123.105:6816/2022540299,v1:192.168.123.105:6817/2022540299]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:24.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:24.979 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:24 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/602465697' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: Detected new or changed devices on vm05 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: Adjusting osd_memory_target on vm05 to 1451M 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: osdmap e43: 8 total, 7 up, 8 in 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: from='osd.7 [v2:192.168.123.105:6816/2022540299,v1:192.168.123.105:6817/2022540299]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:24 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/602465697' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:43:25.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:25.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: Detected new or changed devices on vm05 2026-03-09T17:43:25.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:25.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:25.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:25.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:25.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:43:25.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: Adjusting osd_memory_target on vm05 to 1451M 2026-03-09T17:43:25.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:25.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:25.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:25.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:25.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-09T17:43:25.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: osdmap e43: 8 total, 7 up, 8 in 2026-03-09T17:43:25.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: from='osd.7 [v2:192.168.123.105:6816/2022540299,v1:192.168.123.105:6817/2022540299]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:25.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:25.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-09T17:43:25.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:24 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/602465697' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:43:25.891 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd stat -f json 2026-03-09T17:43:26.062 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:26.180 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:25 vm01 ceph-mon[52793]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-09T17:43:26.180 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:25 vm01 ceph-mon[52793]: osdmap e44: 8 total, 7 up, 8 in 2026-03-09T17:43:26.180 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:25 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:26.180 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:25 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:26.180 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:25 vm01 ceph-mon[52793]: from='osd.7 ' entity='osd.7' 2026-03-09T17:43:26.285 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:43:26.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:25 vm04 ceph-mon[53158]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-09T17:43:26.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:25 vm04 ceph-mon[53158]: osdmap e44: 8 total, 7 up, 8 in 2026-03-09T17:43:26.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:25 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:26.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:25 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:26.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:25 vm04 ceph-mon[53158]: from='osd.7 ' entity='osd.7' 2026-03-09T17:43:26.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:25 vm05 ceph-mon[53831]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-09T17:43:26.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:25 vm05 ceph-mon[53831]: osdmap e44: 8 total, 7 up, 8 in 2026-03-09T17:43:26.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:25 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:26.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:25 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:26.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:25 vm05 ceph-mon[53831]: from='osd.7 ' entity='osd.7' 2026-03-09T17:43:26.320 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:43:25 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-7[67389]: 2026-03-09T17:43:25.824+0000 7f7526852640 -1 osd.7 0 waiting for initial osdmap 2026-03-09T17:43:26.320 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:43:25 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-7[67389]: 2026-03-09T17:43:25.833+0000 7f7521668640 -1 osd.7 44 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:43:26.356 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":44,"num_osds":8,"num_up_osds":7,"osd_up_since":1773078196,"num_in_osds":8,"osd_in_since":1773078195,"num_remapped_pgs":0} 2026-03-09T17:43:27.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:26 vm01 ceph-mon[52793]: purged_snaps scrub starts 2026-03-09T17:43:27.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:26 vm01 ceph-mon[52793]: purged_snaps scrub ok 2026-03-09T17:43:27.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:26 vm01 ceph-mon[52793]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:27.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:26 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:27.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:26 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/69181862' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:43:27.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:26 vm01 ceph-mon[52793]: osd.7 [v2:192.168.123.105:6816/2022540299,v1:192.168.123.105:6817/2022540299] boot 2026-03-09T17:43:27.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:26 vm01 ceph-mon[52793]: osdmap e45: 8 total, 8 up, 8 in 2026-03-09T17:43:27.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:26 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:27.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:26 vm04 ceph-mon[53158]: purged_snaps scrub starts 2026-03-09T17:43:27.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:26 vm04 ceph-mon[53158]: purged_snaps scrub ok 2026-03-09T17:43:27.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:26 vm04 ceph-mon[53158]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:27.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:26 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:27.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:26 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/69181862' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:43:27.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:26 vm04 ceph-mon[53158]: osd.7 [v2:192.168.123.105:6816/2022540299,v1:192.168.123.105:6817/2022540299] boot 2026-03-09T17:43:27.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:26 vm04 ceph-mon[53158]: osdmap e45: 8 total, 8 up, 8 in 2026-03-09T17:43:27.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:26 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:27.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:26 vm05 ceph-mon[53831]: purged_snaps scrub starts 2026-03-09T17:43:27.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:26 vm05 ceph-mon[53831]: purged_snaps scrub ok 2026-03-09T17:43:27.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:26 vm05 ceph-mon[53831]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:43:27.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:26 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:27.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:26 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/69181862' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:43:27.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:26 vm05 ceph-mon[53831]: osd.7 [v2:192.168.123.105:6816/2022540299,v1:192.168.123.105:6817/2022540299] boot 2026-03-09T17:43:27.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:26 vm05 ceph-mon[53831]: osdmap e45: 8 total, 8 up, 8 in 2026-03-09T17:43:27.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:26 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:43:27.358 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd stat -f json 2026-03-09T17:43:27.535 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:27.762 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:43:27.838 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":45,"num_osds":8,"num_up_osds":8,"osd_up_since":1773078206,"num_in_osds":8,"osd_in_since":1773078195,"num_remapped_pgs":0} 2026-03-09T17:43:27.839 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd dump --format=json 2026-03-09T17:43:27.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:27 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1062089551' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:43:28.021 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:28.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:27 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1062089551' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:43:28.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:27 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1062089551' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:43:28.410 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:43:28.410 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":46,"fsid":"01455850-1bdf-11f1-910a-9936d43313cc","created":"2026-03-09T17:41:08.651176+0000","modified":"2026-03-09T17:43:27.891868+0000","last_up_change":"2026-03-09T17:43:26.828173+0000","last_in_change":"2026-03-09T17:43:15.061615+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T17:42:35.411689+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"22","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"b4de0462-8141-43d0-a2b5-00b4ec8c42bb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":26,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3943236347},{"type":"v1","addr":"192.168.123.101:6803","nonce":3943236347}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3943236347},{"type":"v1","addr":"192.168.123.101:6805","nonce":3943236347}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":3943236347},{"type":"v1","addr":"192.168.123.101:6809","nonce":3943236347}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3943236347},{"type":"v1","addr":"192.168.123.101:6807","nonce":3943236347}]},"public_addr":"192.168.123.101:6803/3943236347","cluster_addr":"192.168.123.101:6805/3943236347","heartbeat_back_addr":"192.168.123.101:6809/3943236347","heartbeat_front_addr":"192.168.123.101:6807/3943236347","state":["exists","up"]},{"osd":1,"uuid":"651f912d-fa94-4d83-9b72-bb0b98168d7b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":14,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":2752039663},{"type":"v1","addr":"192.168.123.101:6811","nonce":2752039663}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":2752039663},{"type":"v1","addr":"192.168.123.101:6813","nonce":2752039663}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":2752039663},{"type":"v1","addr":"192.168.123.101:6817","nonce":2752039663}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":2752039663},{"type":"v1","addr":"192.168.123.101:6815","nonce":2752039663}]},"public_addr":"192.168.123.101:6811/2752039663","cluster_addr":"192.168.123.101:6813/2752039663","heartbeat_back_addr":"192.168.123.101:6817/2752039663","heartbeat_front_addr":"192.168.123.101:6815/2752039663","state":["exists","up"]},{"osd":2,"uuid":"9afeb5af-65ac-471d-b3a4-b633e11482f9","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":20,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":3952920925},{"type":"v1","addr":"192.168.123.104:6801","nonce":3952920925}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":3952920925},{"type":"v1","addr":"192.168.123.104:6803","nonce":3952920925}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":3952920925},{"type":"v1","addr":"192.168.123.104:6807","nonce":3952920925}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":3952920925},{"type":"v1","addr":"192.168.123.104:6805","nonce":3952920925}]},"public_addr":"192.168.123.104:6801/3952920925","cluster_addr":"192.168.123.104:6803/3952920925","heartbeat_back_addr":"192.168.123.104:6807/3952920925","heartbeat_front_addr":"192.168.123.104:6805/3952920925","state":["exists","up"]},{"osd":3,"uuid":"69f44bac-cbb9-4564-abdc-37247aa0c954","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":26,"up_thru":36,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6808","nonce":2640539626},{"type":"v1","addr":"192.168.123.104:6809","nonce":2640539626}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6810","nonce":2640539626},{"type":"v1","addr":"192.168.123.104:6811","nonce":2640539626}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6814","nonce":2640539626},{"type":"v1","addr":"192.168.123.104:6815","nonce":2640539626}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6812","nonce":2640539626},{"type":"v1","addr":"192.168.123.104:6813","nonce":2640539626}]},"public_addr":"192.168.123.104:6809/2640539626","cluster_addr":"192.168.123.104:6811/2640539626","heartbeat_back_addr":"192.168.123.104:6815/2640539626","heartbeat_front_addr":"192.168.123.104:6813/2640539626","state":["exists","up"]},{"osd":4,"uuid":"f2ea538a-0dea-48f3-87d2-b20de90a8311","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":31,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6816","nonce":3638700262},{"type":"v1","addr":"192.168.123.104:6817","nonce":3638700262}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6818","nonce":3638700262},{"type":"v1","addr":"192.168.123.104:6819","nonce":3638700262}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6822","nonce":3638700262},{"type":"v1","addr":"192.168.123.104:6823","nonce":3638700262}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6820","nonce":3638700262},{"type":"v1","addr":"192.168.123.104:6821","nonce":3638700262}]},"public_addr":"192.168.123.104:6817/3638700262","cluster_addr":"192.168.123.104:6819/3638700262","heartbeat_back_addr":"192.168.123.104:6823/3638700262","heartbeat_front_addr":"192.168.123.104:6821/3638700262","state":["exists","up"]},{"osd":5,"uuid":"4dcf9f68-3a19-43f6-b608-2d0481914d8a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":35,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":2510326788},{"type":"v1","addr":"192.168.123.105:6801","nonce":2510326788}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":2510326788},{"type":"v1","addr":"192.168.123.105:6803","nonce":2510326788}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":2510326788},{"type":"v1","addr":"192.168.123.105:6807","nonce":2510326788}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":2510326788},{"type":"v1","addr":"192.168.123.105:6805","nonce":2510326788}]},"public_addr":"192.168.123.105:6801/2510326788","cluster_addr":"192.168.123.105:6803/2510326788","heartbeat_back_addr":"192.168.123.105:6807/2510326788","heartbeat_front_addr":"192.168.123.105:6805/2510326788","state":["exists","up"]},{"osd":6,"uuid":"136fb72c-868c-451b-a36d-2e332f552fff","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":41,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6808","nonce":4036098223},{"type":"v1","addr":"192.168.123.105:6809","nonce":4036098223}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6810","nonce":4036098223},{"type":"v1","addr":"192.168.123.105:6811","nonce":4036098223}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6814","nonce":4036098223},{"type":"v1","addr":"192.168.123.105:6815","nonce":4036098223}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6812","nonce":4036098223},{"type":"v1","addr":"192.168.123.105:6813","nonce":4036098223}]},"public_addr":"192.168.123.105:6809/4036098223","cluster_addr":"192.168.123.105:6811/4036098223","heartbeat_back_addr":"192.168.123.105:6815/4036098223","heartbeat_front_addr":"192.168.123.105:6813/4036098223","state":["exists","up"]},{"osd":7,"uuid":"11afd218-abae-4bf5-ac97-bba7cd66648d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":45,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6816","nonce":2022540299},{"type":"v1","addr":"192.168.123.105:6817","nonce":2022540299}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6818","nonce":2022540299},{"type":"v1","addr":"192.168.123.105:6819","nonce":2022540299}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6822","nonce":2022540299},{"type":"v1","addr":"192.168.123.105:6823","nonce":2022540299}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6820","nonce":2022540299},{"type":"v1","addr":"192.168.123.105:6821","nonce":2022540299}]},"public_addr":"192.168.123.105:6817/2022540299","cluster_addr":"192.168.123.105:6819/2022540299","heartbeat_back_addr":"192.168.123.105:6823/2022540299","heartbeat_front_addr":"192.168.123.105:6821/2022540299","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:11.699546+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:21.391777+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:31.666177+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:42.659295+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:53.076638+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:43:03.598469+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:43:13.354383+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:43:24.383876+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:6801/4099002418":"2026-03-10T17:41:31.362732+0000","192.168.123.101:0/1010633484":"2026-03-10T17:41:31.362732+0000","192.168.123.101:0/2206736793":"2026-03-10T17:41:31.362732+0000","192.168.123.101:6800/4099002418":"2026-03-10T17:41:31.362732+0000","192.168.123.101:6801/630561731":"2026-03-10T17:41:20.447130+0000","192.168.123.101:6800/630561731":"2026-03-10T17:41:20.447130+0000","192.168.123.101:0/13233085":"2026-03-10T17:41:20.447130+0000","192.168.123.101:0/702698574":"2026-03-10T17:41:20.447130+0000","192.168.123.101:0/573666992":"2026-03-10T17:41:31.362732+0000","192.168.123.101:0/1188799241":"2026-03-10T17:41:20.447130+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T17:43:28.875 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-09T17:42:35.411689+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '22', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 7.889999866485596, 'score_stable': 7.889999866485596, 'optimal_score': 0.3799999952316284, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-09T17:43:28.875 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd pool get .mgr pg_num 2026-03-09T17:43:29.057 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:29.114 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:28 vm01 ceph-mon[52793]: pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:29.114 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:28 vm01 ceph-mon[52793]: osdmap e46: 8 total, 8 up, 8 in 2026-03-09T17:43:29.114 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:28 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3085192385' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:43:29.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:28 vm04 ceph-mon[53158]: pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:29.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:28 vm04 ceph-mon[53158]: osdmap e46: 8 total, 8 up, 8 in 2026-03-09T17:43:29.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:28 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3085192385' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:43:29.296 INFO:teuthology.orchestra.run.vm01.stdout:pg_num: 1 2026-03-09T17:43:29.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:28 vm05 ceph-mon[53831]: pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:29.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:28 vm05 ceph-mon[53831]: osdmap e46: 8 total, 8 up, 8 in 2026-03-09T17:43:29.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:28 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3085192385' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:43:29.363 INFO:tasks.cephadm:Adding ceph.iscsi.iscsi.a on vm01 2026-03-09T17:43:29.363 INFO:tasks.cephadm:Adding ceph.iscsi.iscsi.b on vm05 2026-03-09T17:43:29.363 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd pool create datapool 3 3 replicated 2026-03-09T17:43:29.545 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.c/config 2026-03-09T17:43:29.909 INFO:teuthology.orchestra.run.vm05.stderr:pool 'datapool' created 2026-03-09T17:43:29.921 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:29 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1268594636' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-09T17:43:29.921 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:29 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/4007921524' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-09T17:43:29.921 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:29 vm05 ceph-mon[53831]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-09T17:43:29.976 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- rbd pool init datapool 2026-03-09T17:43:30.157 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.c/config 2026-03-09T17:43:30.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:29 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1268594636' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-09T17:43:30.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:29 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/4007921524' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-09T17:43:30.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:29 vm01 ceph-mon[52793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-09T17:43:30.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:29 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1268594636' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-09T17:43:30.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:29 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/4007921524' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-09T17:43:30.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:29 vm04 ceph-mon[53158]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-09T17:43:31.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:30 vm01 ceph-mon[52793]: pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:31.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:30 vm01 ceph-mon[52793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-09T17:43:31.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:30 vm01 ceph-mon[52793]: osdmap e47: 8 total, 8 up, 8 in 2026-03-09T17:43:31.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:30 vm01 ceph-mon[52793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-09T17:43:31.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:30 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3382006535' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-09T17:43:31.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:30 vm04 ceph-mon[53158]: pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:31.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:30 vm04 ceph-mon[53158]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-09T17:43:31.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:30 vm04 ceph-mon[53158]: osdmap e47: 8 total, 8 up, 8 in 2026-03-09T17:43:31.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:30 vm04 ceph-mon[53158]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-09T17:43:31.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:30 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3382006535' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-09T17:43:31.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:30 vm05 ceph-mon[53831]: pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:31.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:30 vm05 ceph-mon[53831]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-09T17:43:31.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:30 vm05 ceph-mon[53831]: osdmap e47: 8 total, 8 up, 8 in 2026-03-09T17:43:31.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:30 vm05 ceph-mon[53831]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-09T17:43:31.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:30 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3382006535' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-09T17:43:32.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:31 vm01 ceph-mon[52793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-09T17:43:32.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:31 vm01 ceph-mon[52793]: osdmap e48: 8 total, 8 up, 8 in 2026-03-09T17:43:32.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:31 vm04 ceph-mon[53158]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-09T17:43:32.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:31 vm04 ceph-mon[53158]: osdmap e48: 8 total, 8 up, 8 in 2026-03-09T17:43:32.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:31 vm05 ceph-mon[53831]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-09T17:43:32.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:31 vm05 ceph-mon[53831]: osdmap e48: 8 total, 8 up, 8 in 2026-03-09T17:43:32.991 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph orch apply iscsi datapool admin admin --trusted_ip_list 192.168.123.101,192.168.123.105 --placement '2;vm01=iscsi.a;vm05=iscsi.b' 2026-03-09T17:43:33.175 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.c/config 2026-03-09T17:43:33.198 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:32 vm05 ceph-mon[53831]: pgmap v98: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:33.198 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:32 vm05 ceph-mon[53831]: osdmap e49: 8 total, 8 up, 8 in 2026-03-09T17:43:33.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:32 vm01 ceph-mon[52793]: pgmap v98: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:33.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:32 vm01 ceph-mon[52793]: osdmap e49: 8 total, 8 up, 8 in 2026-03-09T17:43:33.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:32 vm04 ceph-mon[53158]: pgmap v98: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:33.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:32 vm04 ceph-mon[53158]: osdmap e49: 8 total, 8 up, 8 in 2026-03-09T17:43:33.418 INFO:teuthology.orchestra.run.vm05.stdout:Scheduled iscsi.datapool update... 2026-03-09T17:43:33.494 INFO:tasks.cephadm:Distributing iscsi-gateway.cfg... 2026-03-09T17:43:33.494 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:43:33.494 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/iscsi-gateway.cfg 2026-03-09T17:43:33.522 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:43:33.522 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/iscsi-gateway.cfg 2026-03-09T17:43:33.557 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-09T17:43:33.557 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/ceph/iscsi-gateway.cfg 2026-03-09T17:43:33.584 DEBUG:teuthology.orchestra.run.vm01:iscsi.iscsi.a> sudo journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@iscsi.iscsi.a.service 2026-03-09T17:43:33.588 DEBUG:teuthology.orchestra.run.vm05:iscsi.iscsi.b> sudo journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@iscsi.iscsi.b.service 2026-03-09T17:43:33.627 INFO:tasks.cephadm:Setting up client nodes... 2026-03-09T17:43:33.627 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph auth get-or-create client.0 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-09T17:43:33.881 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:34.165 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:33 vm01 ceph-mon[52793]: osdmap e50: 8 total, 8 up, 8 in 2026-03-09T17:43:34.165 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:34.165 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:34.165 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:34.165 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:34.165 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:34.165 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-09T17:43:34.165 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-09T17:43:34.165 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:33 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:34.222 INFO:teuthology.orchestra.run.vm01.stdout:[client.0] 2026-03-09T17:43:34.223 INFO:teuthology.orchestra.run.vm01.stdout: key = AQDGBq9pgJLeDBAAFtEo5f677yD1h5OcEuutSg== 2026-03-09T17:43:34.286 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:43:34.286 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.client.0.keyring 2026-03-09T17:43:34.286 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod 0644 /etc/ceph/ceph.client.0.keyring 2026-03-09T17:43:34.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:33 vm04 ceph-mon[53158]: osdmap e50: 8 total, 8 up, 8 in 2026-03-09T17:43:34.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:33 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:34.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:33 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:34.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:33 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:34.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:33 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:34.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:33 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:34.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:33 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-09T17:43:34.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:33 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-09T17:43:34.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:33 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:34.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:33 vm05 ceph-mon[53831]: osdmap e50: 8 total, 8 up, 8 in 2026-03-09T17:43:34.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:33 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:34.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:33 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:34.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:33 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:34.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:33 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:34.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:33 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:34.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:33 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-09T17:43:34.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:33 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-09T17:43:34.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:33 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:34.334 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph auth get-or-create client.1 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-09T17:43:34.523 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.b/config 2026-03-09T17:43:34.813 INFO:teuthology.orchestra.run.vm04.stdout:[client.1] 2026-03-09T17:43:34.813 INFO:teuthology.orchestra.run.vm04.stdout: key = AQDGBq9pCR8HMBAAMzoDkAceEqnJkJNaow00FA== 2026-03-09T17:43:34.869 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:43:34.869 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/ceph.client.1.keyring 2026-03-09T17:43:34.869 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod 0644 /etc/ceph/ceph.client.1.keyring 2026-03-09T17:43:34.912 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph auth get-or-create client.2 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-09T17:43:35.146 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.c/config 2026-03-09T17:43:35.204 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: pgmap v101: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: from='client.24334 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.101,192.168.123.105", "placement": "2;vm01=iscsi.a;vm05=iscsi.b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: Saving service iscsi.datapool spec with placement vm01=iscsi.a;vm05=iscsi.b;count:2 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: Deploying daemon iscsi.iscsi.a on vm01 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2246281524' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2234029222' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: from='client.? 192.168.123.104:0/3634734415' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: from='client.? 192.168.123.104:0/3634734415' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-09T17:43:35.205 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:34 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3552238165' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/4099002418"}]: dispatch 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: pgmap v101: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: from='client.24334 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.101,192.168.123.105", "placement": "2;vm01=iscsi.a;vm05=iscsi.b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: Saving service iscsi.datapool spec with placement vm01=iscsi.a;vm05=iscsi.b;count:2 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: Deploying daemon iscsi.iscsi.a on vm01 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2246281524' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2234029222' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-09T17:43:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: from='client.? 192.168.123.104:0/3634734415' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:35.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: from='client.? 192.168.123.104:0/3634734415' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-09T17:43:35.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:34 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3552238165' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/4099002418"}]: dispatch 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: pgmap v101: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: from='client.24334 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.101,192.168.123.105", "placement": "2;vm01=iscsi.a;vm05=iscsi.b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: Saving service iscsi.datapool spec with placement vm01=iscsi.a;vm05=iscsi.b;count:2 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: Deploying daemon iscsi.iscsi.a on vm01 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2246281524' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2234029222' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: from='client.? 192.168.123.104:0/3634734415' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: from='client.? 192.168.123.104:0/3634734415' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-09T17:43:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:34 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3552238165' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/4099002418"}]: dispatch 2026-03-09T17:43:35.617 INFO:teuthology.orchestra.run.vm05.stdout:[client.2] 2026-03-09T17:43:35.617 INFO:teuthology.orchestra.run.vm05.stdout: key = AQDHBq9pK9k2JBAAYB1RDLPC7xqO7vAIqjxy2w== 2026-03-09T17:43:35.757 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-09T17:43:35.757 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/ceph/ceph.client.2.keyring 2026-03-09T17:43:35.757 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod 0644 /etc/ceph/ceph.client.2.keyring 2026-03-09T17:43:35.852 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-09T17:43:35.852 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-09T17:43:35.852 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph mgr dump --format=json 2026-03-09T17:43:36.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:35 vm05 ceph-mon[53831]: Deploying daemon iscsi.iscsi.b on vm05 2026-03-09T17:43:36.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:35 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:36.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:35 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:36.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:35 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:36.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:35 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:36.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:35 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:36.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:35 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1899859374' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:36.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:35 vm05 ceph-mon[53831]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:36.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:35 vm05 ceph-mon[53831]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-09T17:43:36.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:35 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3552238165' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/4099002418"}]': finished 2026-03-09T17:43:36.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:35 vm05 ceph-mon[53831]: osdmap e51: 8 total, 8 up, 8 in 2026-03-09T17:43:36.133 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:36.163 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:35 vm01 ceph-mon[52793]: Deploying daemon iscsi.iscsi.b on vm05 2026-03-09T17:43:36.163 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:35 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:36.163 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:35 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:36.163 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:35 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:36.163 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:35 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:36.163 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:35 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:36.163 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:35 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1899859374' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:36.163 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:35 vm01 ceph-mon[52793]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:36.163 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:35 vm01 ceph-mon[52793]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-09T17:43:36.163 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:35 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3552238165' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/4099002418"}]': finished 2026-03-09T17:43:36.163 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:35 vm01 ceph-mon[52793]: osdmap e51: 8 total, 8 up, 8 in 2026-03-09T17:43:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:35 vm04 ceph-mon[53158]: Deploying daemon iscsi.iscsi.b on vm05 2026-03-09T17:43:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:35 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:35 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:35 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:35 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:35 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:35 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1899859374' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:35 vm04 ceph-mon[53158]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:43:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:35 vm04 ceph-mon[53158]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-09T17:43:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:35 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3552238165' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/4099002418"}]': finished 2026-03-09T17:43:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:35 vm04 ceph-mon[53158]: osdmap e51: 8 total, 8 up, 8 in 2026-03-09T17:43:36.434 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:43:36.506 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":14,"flags":0,"active_gid":14150,"active_name":"x","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6800","nonce":792002087},{"type":"v1","addr":"192.168.123.101:6801","nonce":792002087}]},"active_addr":"192.168.123.101:6801/792002087","active_change":"2026-03-09T17:41:31.363114+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.101:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":1358501084}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":413698299}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":1332728170}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":3179920097}]}]} 2026-03-09T17:43:36.507 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-09T17:43:36.507 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-09T17:43:36.507 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd dump --format=json 2026-03-09T17:43:36.695 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:36.927 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:43:36.927 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":51,"fsid":"01455850-1bdf-11f1-910a-9936d43313cc","created":"2026-03-09T17:41:08.651176+0000","modified":"2026-03-09T17:43:35.808858+0000","last_up_change":"2026-03-09T17:43:26.828173+0000","last_in_change":"2026-03-09T17:43:15.061615+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":2,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T17:42:35.411689+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"22","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":"datapool","create_time":"2026-03-09T17:43:29.794454+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":3,"pg_placement_num":3,"pg_placement_num_target":3,"pg_num_target":3,"pg_num_pending":3,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"50","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":50,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":5.3299999237060547,"score_stable":5.3299999237060547,"optimal_score":0.75,"raw_score_acting":4,"raw_score_stable":4,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"b4de0462-8141-43d0-a2b5-00b4ec8c42bb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":26,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3943236347},{"type":"v1","addr":"192.168.123.101:6803","nonce":3943236347}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3943236347},{"type":"v1","addr":"192.168.123.101:6805","nonce":3943236347}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":3943236347},{"type":"v1","addr":"192.168.123.101:6809","nonce":3943236347}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3943236347},{"type":"v1","addr":"192.168.123.101:6807","nonce":3943236347}]},"public_addr":"192.168.123.101:6803/3943236347","cluster_addr":"192.168.123.101:6805/3943236347","heartbeat_back_addr":"192.168.123.101:6809/3943236347","heartbeat_front_addr":"192.168.123.101:6807/3943236347","state":["exists","up"]},{"osd":1,"uuid":"651f912d-fa94-4d83-9b72-bb0b98168d7b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":14,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":2752039663},{"type":"v1","addr":"192.168.123.101:6811","nonce":2752039663}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":2752039663},{"type":"v1","addr":"192.168.123.101:6813","nonce":2752039663}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":2752039663},{"type":"v1","addr":"192.168.123.101:6817","nonce":2752039663}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":2752039663},{"type":"v1","addr":"192.168.123.101:6815","nonce":2752039663}]},"public_addr":"192.168.123.101:6811/2752039663","cluster_addr":"192.168.123.101:6813/2752039663","heartbeat_back_addr":"192.168.123.101:6817/2752039663","heartbeat_front_addr":"192.168.123.101:6815/2752039663","state":["exists","up"]},{"osd":2,"uuid":"9afeb5af-65ac-471d-b3a4-b633e11482f9","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":47,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":3952920925},{"type":"v1","addr":"192.168.123.104:6801","nonce":3952920925}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":3952920925},{"type":"v1","addr":"192.168.123.104:6803","nonce":3952920925}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":3952920925},{"type":"v1","addr":"192.168.123.104:6807","nonce":3952920925}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":3952920925},{"type":"v1","addr":"192.168.123.104:6805","nonce":3952920925}]},"public_addr":"192.168.123.104:6801/3952920925","cluster_addr":"192.168.123.104:6803/3952920925","heartbeat_back_addr":"192.168.123.104:6807/3952920925","heartbeat_front_addr":"192.168.123.104:6805/3952920925","state":["exists","up"]},{"osd":3,"uuid":"69f44bac-cbb9-4564-abdc-37247aa0c954","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":26,"up_thru":47,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6808","nonce":2640539626},{"type":"v1","addr":"192.168.123.104:6809","nonce":2640539626}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6810","nonce":2640539626},{"type":"v1","addr":"192.168.123.104:6811","nonce":2640539626}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6814","nonce":2640539626},{"type":"v1","addr":"192.168.123.104:6815","nonce":2640539626}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6812","nonce":2640539626},{"type":"v1","addr":"192.168.123.104:6813","nonce":2640539626}]},"public_addr":"192.168.123.104:6809/2640539626","cluster_addr":"192.168.123.104:6811/2640539626","heartbeat_back_addr":"192.168.123.104:6815/2640539626","heartbeat_front_addr":"192.168.123.104:6813/2640539626","state":["exists","up"]},{"osd":4,"uuid":"f2ea538a-0dea-48f3-87d2-b20de90a8311","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":31,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6816","nonce":3638700262},{"type":"v1","addr":"192.168.123.104:6817","nonce":3638700262}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6818","nonce":3638700262},{"type":"v1","addr":"192.168.123.104:6819","nonce":3638700262}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6822","nonce":3638700262},{"type":"v1","addr":"192.168.123.104:6823","nonce":3638700262}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6820","nonce":3638700262},{"type":"v1","addr":"192.168.123.104:6821","nonce":3638700262}]},"public_addr":"192.168.123.104:6817/3638700262","cluster_addr":"192.168.123.104:6819/3638700262","heartbeat_back_addr":"192.168.123.104:6823/3638700262","heartbeat_front_addr":"192.168.123.104:6821/3638700262","state":["exists","up"]},{"osd":5,"uuid":"4dcf9f68-3a19-43f6-b608-2d0481914d8a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":35,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":2510326788},{"type":"v1","addr":"192.168.123.105:6801","nonce":2510326788}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":2510326788},{"type":"v1","addr":"192.168.123.105:6803","nonce":2510326788}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":2510326788},{"type":"v1","addr":"192.168.123.105:6807","nonce":2510326788}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":2510326788},{"type":"v1","addr":"192.168.123.105:6805","nonce":2510326788}]},"public_addr":"192.168.123.105:6801/2510326788","cluster_addr":"192.168.123.105:6803/2510326788","heartbeat_back_addr":"192.168.123.105:6807/2510326788","heartbeat_front_addr":"192.168.123.105:6805/2510326788","state":["exists","up"]},{"osd":6,"uuid":"136fb72c-868c-451b-a36d-2e332f552fff","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":41,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6808","nonce":4036098223},{"type":"v1","addr":"192.168.123.105:6809","nonce":4036098223}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6810","nonce":4036098223},{"type":"v1","addr":"192.168.123.105:6811","nonce":4036098223}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6814","nonce":4036098223},{"type":"v1","addr":"192.168.123.105:6815","nonce":4036098223}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6812","nonce":4036098223},{"type":"v1","addr":"192.168.123.105:6813","nonce":4036098223}]},"public_addr":"192.168.123.105:6809/4036098223","cluster_addr":"192.168.123.105:6811/4036098223","heartbeat_back_addr":"192.168.123.105:6815/4036098223","heartbeat_front_addr":"192.168.123.105:6813/4036098223","state":["exists","up"]},{"osd":7,"uuid":"11afd218-abae-4bf5-ac97-bba7cd66648d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":45,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6816","nonce":2022540299},{"type":"v1","addr":"192.168.123.105:6817","nonce":2022540299}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6818","nonce":2022540299},{"type":"v1","addr":"192.168.123.105:6819","nonce":2022540299}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6822","nonce":2022540299},{"type":"v1","addr":"192.168.123.105:6823","nonce":2022540299}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6820","nonce":2022540299},{"type":"v1","addr":"192.168.123.105:6821","nonce":2022540299}]},"public_addr":"192.168.123.105:6817/2022540299","cluster_addr":"192.168.123.105:6819/2022540299","heartbeat_back_addr":"192.168.123.105:6823/2022540299","heartbeat_front_addr":"192.168.123.105:6821/2022540299","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:11.699546+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:21.391777+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:31.666177+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:42.659295+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:53.076638+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:43:03.598469+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:43:13.354383+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:43:24.383876+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:0/1010633484":"2026-03-10T17:41:31.362732+0000","192.168.123.101:0/2206736793":"2026-03-10T17:41:31.362732+0000","192.168.123.101:6800/4099002418":"2026-03-10T17:41:31.362732+0000","192.168.123.101:6801/630561731":"2026-03-10T17:41:20.447130+0000","192.168.123.101:6800/630561731":"2026-03-10T17:41:20.447130+0000","192.168.123.101:0/13233085":"2026-03-10T17:41:20.447130+0000","192.168.123.101:0/702698574":"2026-03-10T17:41:20.447130+0000","192.168.123.101:0/573666992":"2026-03-10T17:41:31.362732+0000","192.168.123.101:0/1188799241":"2026-03-10T17:41:20.447130+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T17:43:36.973 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: pgmap v102: 4 pgs: 2 creating+peering, 2 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:37.013 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-09T17:43:37.013 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd dump --format=json 2026-03-09T17:43:37.208 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: Checking pool "datapool" exists for service iscsi.datapool 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/441926221' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/844666508' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1010633484"}]: dispatch 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1279101985' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm01"}]: dispatch 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm05"}]: dispatch 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.229 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:36 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3026399740' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: pgmap v102: 4 pgs: 2 creating+peering, 2 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: Checking pool "datapool" exists for service iscsi.datapool 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/441926221' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/844666508' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1010633484"}]: dispatch 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1279101985' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm01"}]: dispatch 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm05"}]: dispatch 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:36 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3026399740' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: pgmap v102: 4 pgs: 2 creating+peering, 2 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: Checking pool "datapool" exists for service iscsi.datapool 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/441926221' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/844666508' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1010633484"}]: dispatch 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1279101985' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm01"}]: dispatch 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm05"}]: dispatch 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:43:37.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:36 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3026399740' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:43:37.434 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:43:37.434 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":52,"fsid":"01455850-1bdf-11f1-910a-9936d43313cc","created":"2026-03-09T17:41:08.651176+0000","modified":"2026-03-09T17:43:36.951590+0000","last_up_change":"2026-03-09T17:43:26.828173+0000","last_in_change":"2026-03-09T17:43:15.061615+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":2,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T17:42:35.411689+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"22","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":"datapool","create_time":"2026-03-09T17:43:29.794454+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":3,"pg_placement_num":3,"pg_placement_num_target":3,"pg_num_target":3,"pg_num_pending":3,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"50","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":50,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":5.3299999237060547,"score_stable":5.3299999237060547,"optimal_score":0.75,"raw_score_acting":4,"raw_score_stable":4,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"b4de0462-8141-43d0-a2b5-00b4ec8c42bb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":26,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3943236347},{"type":"v1","addr":"192.168.123.101:6803","nonce":3943236347}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3943236347},{"type":"v1","addr":"192.168.123.101:6805","nonce":3943236347}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":3943236347},{"type":"v1","addr":"192.168.123.101:6809","nonce":3943236347}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3943236347},{"type":"v1","addr":"192.168.123.101:6807","nonce":3943236347}]},"public_addr":"192.168.123.101:6803/3943236347","cluster_addr":"192.168.123.101:6805/3943236347","heartbeat_back_addr":"192.168.123.101:6809/3943236347","heartbeat_front_addr":"192.168.123.101:6807/3943236347","state":["exists","up"]},{"osd":1,"uuid":"651f912d-fa94-4d83-9b72-bb0b98168d7b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":14,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":2752039663},{"type":"v1","addr":"192.168.123.101:6811","nonce":2752039663}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":2752039663},{"type":"v1","addr":"192.168.123.101:6813","nonce":2752039663}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":2752039663},{"type":"v1","addr":"192.168.123.101:6817","nonce":2752039663}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":2752039663},{"type":"v1","addr":"192.168.123.101:6815","nonce":2752039663}]},"public_addr":"192.168.123.101:6811/2752039663","cluster_addr":"192.168.123.101:6813/2752039663","heartbeat_back_addr":"192.168.123.101:6817/2752039663","heartbeat_front_addr":"192.168.123.101:6815/2752039663","state":["exists","up"]},{"osd":2,"uuid":"9afeb5af-65ac-471d-b3a4-b633e11482f9","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":47,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":3952920925},{"type":"v1","addr":"192.168.123.104:6801","nonce":3952920925}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":3952920925},{"type":"v1","addr":"192.168.123.104:6803","nonce":3952920925}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":3952920925},{"type":"v1","addr":"192.168.123.104:6807","nonce":3952920925}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":3952920925},{"type":"v1","addr":"192.168.123.104:6805","nonce":3952920925}]},"public_addr":"192.168.123.104:6801/3952920925","cluster_addr":"192.168.123.104:6803/3952920925","heartbeat_back_addr":"192.168.123.104:6807/3952920925","heartbeat_front_addr":"192.168.123.104:6805/3952920925","state":["exists","up"]},{"osd":3,"uuid":"69f44bac-cbb9-4564-abdc-37247aa0c954","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":26,"up_thru":47,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6808","nonce":2640539626},{"type":"v1","addr":"192.168.123.104:6809","nonce":2640539626}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6810","nonce":2640539626},{"type":"v1","addr":"192.168.123.104:6811","nonce":2640539626}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6814","nonce":2640539626},{"type":"v1","addr":"192.168.123.104:6815","nonce":2640539626}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6812","nonce":2640539626},{"type":"v1","addr":"192.168.123.104:6813","nonce":2640539626}]},"public_addr":"192.168.123.104:6809/2640539626","cluster_addr":"192.168.123.104:6811/2640539626","heartbeat_back_addr":"192.168.123.104:6815/2640539626","heartbeat_front_addr":"192.168.123.104:6813/2640539626","state":["exists","up"]},{"osd":4,"uuid":"f2ea538a-0dea-48f3-87d2-b20de90a8311","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":31,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6816","nonce":3638700262},{"type":"v1","addr":"192.168.123.104:6817","nonce":3638700262}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6818","nonce":3638700262},{"type":"v1","addr":"192.168.123.104:6819","nonce":3638700262}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6822","nonce":3638700262},{"type":"v1","addr":"192.168.123.104:6823","nonce":3638700262}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6820","nonce":3638700262},{"type":"v1","addr":"192.168.123.104:6821","nonce":3638700262}]},"public_addr":"192.168.123.104:6817/3638700262","cluster_addr":"192.168.123.104:6819/3638700262","heartbeat_back_addr":"192.168.123.104:6823/3638700262","heartbeat_front_addr":"192.168.123.104:6821/3638700262","state":["exists","up"]},{"osd":5,"uuid":"4dcf9f68-3a19-43f6-b608-2d0481914d8a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":35,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":2510326788},{"type":"v1","addr":"192.168.123.105:6801","nonce":2510326788}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":2510326788},{"type":"v1","addr":"192.168.123.105:6803","nonce":2510326788}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":2510326788},{"type":"v1","addr":"192.168.123.105:6807","nonce":2510326788}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":2510326788},{"type":"v1","addr":"192.168.123.105:6805","nonce":2510326788}]},"public_addr":"192.168.123.105:6801/2510326788","cluster_addr":"192.168.123.105:6803/2510326788","heartbeat_back_addr":"192.168.123.105:6807/2510326788","heartbeat_front_addr":"192.168.123.105:6805/2510326788","state":["exists","up"]},{"osd":6,"uuid":"136fb72c-868c-451b-a36d-2e332f552fff","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":41,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6808","nonce":4036098223},{"type":"v1","addr":"192.168.123.105:6809","nonce":4036098223}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6810","nonce":4036098223},{"type":"v1","addr":"192.168.123.105:6811","nonce":4036098223}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6814","nonce":4036098223},{"type":"v1","addr":"192.168.123.105:6815","nonce":4036098223}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6812","nonce":4036098223},{"type":"v1","addr":"192.168.123.105:6813","nonce":4036098223}]},"public_addr":"192.168.123.105:6809/4036098223","cluster_addr":"192.168.123.105:6811/4036098223","heartbeat_back_addr":"192.168.123.105:6815/4036098223","heartbeat_front_addr":"192.168.123.105:6813/4036098223","state":["exists","up"]},{"osd":7,"uuid":"11afd218-abae-4bf5-ac97-bba7cd66648d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":45,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6816","nonce":2022540299},{"type":"v1","addr":"192.168.123.105:6817","nonce":2022540299}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6818","nonce":2022540299},{"type":"v1","addr":"192.168.123.105:6819","nonce":2022540299}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6822","nonce":2022540299},{"type":"v1","addr":"192.168.123.105:6823","nonce":2022540299}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6820","nonce":2022540299},{"type":"v1","addr":"192.168.123.105:6821","nonce":2022540299}]},"public_addr":"192.168.123.105:6817/2022540299","cluster_addr":"192.168.123.105:6819/2022540299","heartbeat_back_addr":"192.168.123.105:6823/2022540299","heartbeat_front_addr":"192.168.123.105:6821/2022540299","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:11.699546+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:21.391777+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:31.666177+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:42.659295+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:42:53.076638+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:43:03.598469+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:43:13.354383+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:43:24.383876+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:0/2206736793":"2026-03-10T17:41:31.362732+0000","192.168.123.101:6800/4099002418":"2026-03-10T17:41:31.362732+0000","192.168.123.101:6801/630561731":"2026-03-10T17:41:20.447130+0000","192.168.123.101:6800/630561731":"2026-03-10T17:41:20.447130+0000","192.168.123.101:0/13233085":"2026-03-10T17:41:20.447130+0000","192.168.123.101:0/702698574":"2026-03-10T17:41:20.447130+0000","192.168.123.101:0/573666992":"2026-03-10T17:41:31.362732+0000","192.168.123.101:0/1188799241":"2026-03-10T17:41:20.447130+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T17:43:37.487 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph tell osd.0 flush_pg_stats 2026-03-09T17:43:37.487 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph tell osd.1 flush_pg_stats 2026-03-09T17:43:37.487 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph tell osd.2 flush_pg_stats 2026-03-09T17:43:37.487 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph tell osd.3 flush_pg_stats 2026-03-09T17:43:37.487 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph tell osd.4 flush_pg_stats 2026-03-09T17:43:37.487 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph tell osd.5 flush_pg_stats 2026-03-09T17:43:37.487 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph tell osd.6 flush_pg_stats 2026-03-09T17:43:37.487 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph tell osd.7 flush_pg_stats 2026-03-09T17:43:37.973 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:37 vm01 ceph-mon[52793]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-09T17:43:37.973 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:37 vm01 ceph-mon[52793]: Adding iSCSI gateway http://:@192.168.123.101:5000 to Dashboard 2026-03-09T17:43:37.973 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:37 vm01 ceph-mon[52793]: Adding iSCSI gateway http://:@192.168.123.105:5000 to Dashboard 2026-03-09T17:43:37.973 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:37 vm01 ceph-mon[52793]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-09T17:43:37.973 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:37 vm01 ceph-mon[52793]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm01"}]: dispatch 2026-03-09T17:43:37.973 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:37 vm01 ceph-mon[52793]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm05"}]: dispatch 2026-03-09T17:43:37.973 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:37 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/844666508' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1010633484"}]': finished 2026-03-09T17:43:37.973 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:37 vm01 ceph-mon[52793]: osdmap e52: 8 total, 8 up, 8 in 2026-03-09T17:43:37.973 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:37 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2082016807' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2206736793"}]: dispatch 2026-03-09T17:43:37.973 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:37 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3770354273' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:43:38.241 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:38 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Removing blocklisted entry for this host : 192.168.123.101:6800/4099002418 2026-03-09T17:43:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:37 vm04 ceph-mon[53158]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-09T17:43:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:37 vm04 ceph-mon[53158]: Adding iSCSI gateway http://:@192.168.123.101:5000 to Dashboard 2026-03-09T17:43:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:37 vm04 ceph-mon[53158]: Adding iSCSI gateway http://:@192.168.123.105:5000 to Dashboard 2026-03-09T17:43:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:37 vm04 ceph-mon[53158]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-09T17:43:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:37 vm04 ceph-mon[53158]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm01"}]: dispatch 2026-03-09T17:43:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:37 vm04 ceph-mon[53158]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm05"}]: dispatch 2026-03-09T17:43:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:37 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/844666508' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1010633484"}]': finished 2026-03-09T17:43:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:37 vm04 ceph-mon[53158]: osdmap e52: 8 total, 8 up, 8 in 2026-03-09T17:43:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:37 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2082016807' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2206736793"}]: dispatch 2026-03-09T17:43:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:37 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3770354273' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:43:38.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:37 vm05 ceph-mon[53831]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-09T17:43:38.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:37 vm05 ceph-mon[53831]: Adding iSCSI gateway http://:@192.168.123.101:5000 to Dashboard 2026-03-09T17:43:38.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:37 vm05 ceph-mon[53831]: Adding iSCSI gateway http://:@192.168.123.105:5000 to Dashboard 2026-03-09T17:43:38.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:37 vm05 ceph-mon[53831]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-09T17:43:38.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:37 vm05 ceph-mon[53831]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm01"}]: dispatch 2026-03-09T17:43:38.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:37 vm05 ceph-mon[53831]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm05"}]: dispatch 2026-03-09T17:43:38.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:37 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/844666508' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1010633484"}]': finished 2026-03-09T17:43:38.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:37 vm05 ceph-mon[53831]: osdmap e52: 8 total, 8 up, 8 in 2026-03-09T17:43:38.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:37 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2082016807' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2206736793"}]: dispatch 2026-03-09T17:43:38.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:37 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3770354273' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:43:38.397 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:38.419 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:38.420 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:38.434 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:38.467 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:38.478 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:38.478 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:38.524 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:39.139 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Successfully removed blocklist entry 2026-03-09T17:43:39.140 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Removing blocklisted entry for this host : 192.168.123.101:6801/630561731 2026-03-09T17:43:39.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:38 vm01 ceph-mon[52793]: pgmap v105: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 374 B/s rd, 562 B/s wr, 2 op/s 2026-03-09T17:43:39.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:38 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2082016807' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2206736793"}]': finished 2026-03-09T17:43:39.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:38 vm01 ceph-mon[52793]: mgrmap e15: x(active, since 2m) 2026-03-09T17:43:39.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:38 vm01 ceph-mon[52793]: osdmap e53: 8 total, 8 up, 8 in 2026-03-09T17:43:39.143 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:38 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/332389395' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/4099002418"}]: dispatch 2026-03-09T17:43:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:38 vm04 ceph-mon[53158]: pgmap v105: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 374 B/s rd, 562 B/s wr, 2 op/s 2026-03-09T17:43:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:38 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2082016807' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2206736793"}]': finished 2026-03-09T17:43:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:38 vm04 ceph-mon[53158]: mgrmap e15: x(active, since 2m) 2026-03-09T17:43:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:38 vm04 ceph-mon[53158]: osdmap e53: 8 total, 8 up, 8 in 2026-03-09T17:43:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:38 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/332389395' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/4099002418"}]: dispatch 2026-03-09T17:43:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:38 vm05 ceph-mon[53831]: pgmap v105: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 374 B/s rd, 562 B/s wr, 2 op/s 2026-03-09T17:43:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:38 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2082016807' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2206736793"}]': finished 2026-03-09T17:43:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:38 vm05 ceph-mon[53831]: mgrmap e15: x(active, since 2m) 2026-03-09T17:43:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:38 vm05 ceph-mon[53831]: osdmap e53: 8 total, 8 up, 8 in 2026-03-09T17:43:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:38 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/332389395' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/4099002418"}]: dispatch 2026-03-09T17:43:39.490 INFO:teuthology.orchestra.run.vm01.stdout:111669149708 2026-03-09T17:43:39.490 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd last-stat-seq osd.3 2026-03-09T17:43:39.583 INFO:teuthology.orchestra.run.vm01.stdout:38654705682 2026-03-09T17:43:39.583 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd last-stat-seq osd.0 2026-03-09T17:43:39.642 INFO:teuthology.orchestra.run.vm01.stdout:193273528324 2026-03-09T17:43:39.642 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd last-stat-seq osd.7 2026-03-09T17:43:39.667 INFO:teuthology.orchestra.run.vm01.stdout:60129542161 2026-03-09T17:43:39.667 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd last-stat-seq osd.1 2026-03-09T17:43:39.670 INFO:teuthology.orchestra.run.vm01.stdout:150323855368 2026-03-09T17:43:39.670 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd last-stat-seq osd.5 2026-03-09T17:43:39.683 INFO:teuthology.orchestra.run.vm01.stdout:133143986186 2026-03-09T17:43:39.683 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd last-stat-seq osd.4 2026-03-09T17:43:39.688 INFO:teuthology.orchestra.run.vm01.stdout:176093659142 2026-03-09T17:43:39.688 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd last-stat-seq osd.6 2026-03-09T17:43:39.700 INFO:teuthology.orchestra.run.vm01.stdout:81604378638 2026-03-09T17:43:39.700 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd last-stat-seq osd.2 2026-03-09T17:43:39.980 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Successfully removed blocklist entry 2026-03-09T17:43:39.980 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Removing blocklisted entry for this host : 192.168.123.101:6800/630561731 2026-03-09T17:43:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:40 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/332389395' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/4099002418"}]': finished 2026-03-09T17:43:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:40 vm04 ceph-mon[53158]: osdmap e54: 8 total, 8 up, 8 in 2026-03-09T17:43:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:40 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3452474675' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/630561731"}]: dispatch 2026-03-09T17:43:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:40 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3452474675' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/630561731"}]': finished 2026-03-09T17:43:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:40 vm04 ceph-mon[53158]: osdmap e55: 8 total, 8 up, 8 in 2026-03-09T17:43:40.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:40 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/332389395' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/4099002418"}]': finished 2026-03-09T17:43:40.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:40 vm05 ceph-mon[53831]: osdmap e54: 8 total, 8 up, 8 in 2026-03-09T17:43:40.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:40 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3452474675' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/630561731"}]: dispatch 2026-03-09T17:43:40.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:40 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3452474675' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/630561731"}]': finished 2026-03-09T17:43:40.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:40 vm05 ceph-mon[53831]: osdmap e55: 8 total, 8 up, 8 in 2026-03-09T17:43:40.343 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:40.431 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:40.448 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:40 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/332389395' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/4099002418"}]': finished 2026-03-09T17:43:40.448 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:40 vm01 ceph-mon[52793]: osdmap e54: 8 total, 8 up, 8 in 2026-03-09T17:43:40.448 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:40 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3452474675' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/630561731"}]: dispatch 2026-03-09T17:43:40.448 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:40 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3452474675' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/630561731"}]': finished 2026-03-09T17:43:40.448 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:40 vm01 ceph-mon[52793]: osdmap e55: 8 total, 8 up, 8 in 2026-03-09T17:43:40.653 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:40.657 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:40.675 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:40.744 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:40.840 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:40.981 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:41.082 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:41 vm01 ceph-mon[52793]: pgmap v108: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 767 B/s wr, 5 op/s 2026-03-09T17:43:41.083 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:41 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2079882460' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/630561731"}]: dispatch 2026-03-09T17:43:41.083 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Successfully removed blocklist entry 2026-03-09T17:43:41.083 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Removing blocklisted entry for this host : 192.168.123.101:0/13233085 2026-03-09T17:43:41.289 INFO:teuthology.orchestra.run.vm01.stdout:193273528324 2026-03-09T17:43:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:41 vm04 ceph-mon[53158]: pgmap v108: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 767 B/s wr, 5 op/s 2026-03-09T17:43:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:41 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2079882460' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/630561731"}]: dispatch 2026-03-09T17:43:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:41 vm05 ceph-mon[53831]: pgmap v108: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 767 B/s wr, 5 op/s 2026-03-09T17:43:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:41 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2079882460' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/630561731"}]: dispatch 2026-03-09T17:43:41.459 INFO:tasks.cephadm.ceph_manager.ceph:need seq 193273528324 got 193273528324 for osd.7 2026-03-09T17:43:41.459 DEBUG:teuthology.parallel:result is None 2026-03-09T17:43:41.467 INFO:teuthology.orchestra.run.vm01.stdout:38654705682 2026-03-09T17:43:41.569 INFO:teuthology.orchestra.run.vm01.stdout:133143986185 2026-03-09T17:43:41.696 INFO:tasks.cephadm.ceph_manager.ceph:need seq 133143986186 got 133143986185 for osd.4 2026-03-09T17:43:41.719 INFO:tasks.cephadm.ceph_manager.ceph:need seq 38654705682 got 38654705682 for osd.0 2026-03-09T17:43:41.720 DEBUG:teuthology.parallel:result is None 2026-03-09T17:43:41.768 INFO:teuthology.orchestra.run.vm01.stdout:81604378637 2026-03-09T17:43:41.838 INFO:tasks.cephadm.ceph_manager.ceph:need seq 81604378638 got 81604378637 for osd.2 2026-03-09T17:43:41.849 INFO:teuthology.orchestra.run.vm01.stdout:60129542160 2026-03-09T17:43:41.854 INFO:teuthology.orchestra.run.vm01.stdout:176093659141 2026-03-09T17:43:41.854 INFO:teuthology.orchestra.run.vm01.stdout:150323855368 2026-03-09T17:43:41.941 INFO:tasks.cephadm.ceph_manager.ceph:need seq 150323855368 got 150323855368 for osd.5 2026-03-09T17:43:41.941 DEBUG:teuthology.parallel:result is None 2026-03-09T17:43:41.983 INFO:tasks.cephadm.ceph_manager.ceph:need seq 60129542161 got 60129542160 for osd.1 2026-03-09T17:43:41.992 INFO:teuthology.orchestra.run.vm01.stdout:111669149708 2026-03-09T17:43:42.015 INFO:tasks.cephadm.ceph_manager.ceph:need seq 176093659142 got 176093659141 for osd.6 2026-03-09T17:43:42.061 INFO:tasks.cephadm.ceph_manager.ceph:need seq 111669149708 got 111669149708 for osd.3 2026-03-09T17:43:42.061 DEBUG:teuthology.parallel:result is None 2026-03-09T17:43:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2079882460' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/630561731"}]': finished 2026-03-09T17:43:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:42 vm01 ceph-mon[52793]: osdmap e56: 8 total, 8 up, 8 in 2026-03-09T17:43:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2969458160' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-09T17:43:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/444886212' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/13233085"}]: dispatch 2026-03-09T17:43:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2194583921' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T17:43:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1221412485' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T17:43:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3251005836' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T17:43:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1811735616' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T17:43:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3927257112' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-09T17:43:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1547665072' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-09T17:43:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3402705527' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-09T17:43:42.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Successfully removed blocklist entry 2026-03-09T17:43:42.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Removing blocklisted entry for this host : 192.168.123.101:0/702698574 2026-03-09T17:43:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2079882460' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/630561731"}]': finished 2026-03-09T17:43:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:42 vm04 ceph-mon[53158]: osdmap e56: 8 total, 8 up, 8 in 2026-03-09T17:43:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2969458160' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-09T17:43:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/444886212' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/13233085"}]: dispatch 2026-03-09T17:43:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2194583921' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T17:43:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1221412485' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T17:43:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3251005836' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T17:43:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1811735616' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T17:43:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3927257112' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-09T17:43:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1547665072' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-09T17:43:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3402705527' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-09T17:43:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2079882460' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/630561731"}]': finished 2026-03-09T17:43:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:42 vm05 ceph-mon[53831]: osdmap e56: 8 total, 8 up, 8 in 2026-03-09T17:43:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2969458160' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-09T17:43:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/444886212' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/13233085"}]: dispatch 2026-03-09T17:43:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2194583921' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T17:43:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1221412485' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T17:43:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3251005836' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T17:43:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1811735616' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T17:43:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3927257112' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-09T17:43:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1547665072' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-09T17:43:42.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3402705527' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-09T17:43:42.697 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd last-stat-seq osd.4 2026-03-09T17:43:42.838 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd last-stat-seq osd.2 2026-03-09T17:43:42.871 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:42.983 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd last-stat-seq osd.1 2026-03-09T17:43:43.015 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph osd last-stat-seq osd.6 2026-03-09T17:43:43.117 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Successfully removed blocklist entry 2026-03-09T17:43:43.117 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Removing blocklisted entry for this host : 192.168.123.101:0/573666992 2026-03-09T17:43:43.122 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:43 vm01 ceph-mon[52793]: pgmap v111: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:43:43.122 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:43 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/444886212' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/13233085"}]': finished 2026-03-09T17:43:43.122 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:43 vm01 ceph-mon[52793]: osdmap e57: 8 total, 8 up, 8 in 2026-03-09T17:43:43.122 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:43 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1828690833' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/702698574"}]: dispatch 2026-03-09T17:43:43.122 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:43 vm01 ceph-mon[52793]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/702698574"}]: dispatch 2026-03-09T17:43:43.170 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:43.221 INFO:teuthology.orchestra.run.vm01.stdout:133143986187 2026-03-09T17:43:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:43 vm04 ceph-mon[53158]: pgmap v111: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:43:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:43 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/444886212' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/13233085"}]': finished 2026-03-09T17:43:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:43 vm04 ceph-mon[53158]: osdmap e57: 8 total, 8 up, 8 in 2026-03-09T17:43:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:43 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1828690833' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/702698574"}]: dispatch 2026-03-09T17:43:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:43 vm04 ceph-mon[53158]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/702698574"}]: dispatch 2026-03-09T17:43:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:43 vm05 ceph-mon[53831]: pgmap v111: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:43:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:43 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/444886212' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/13233085"}]': finished 2026-03-09T17:43:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:43 vm05 ceph-mon[53831]: osdmap e57: 8 total, 8 up, 8 in 2026-03-09T17:43:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:43 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1828690833' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/702698574"}]: dispatch 2026-03-09T17:43:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:43 vm05 ceph-mon[53831]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/702698574"}]: dispatch 2026-03-09T17:43:43.362 INFO:tasks.cephadm.ceph_manager.ceph:need seq 133143986186 got 133143986187 for osd.4 2026-03-09T17:43:43.362 DEBUG:teuthology.parallel:result is None 2026-03-09T17:43:43.543 INFO:teuthology.orchestra.run.vm01.stdout:81604378639 2026-03-09T17:43:43.559 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:43.602 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:43.647 INFO:tasks.cephadm.ceph_manager.ceph:need seq 81604378638 got 81604378639 for osd.2 2026-03-09T17:43:43.647 DEBUG:teuthology.parallel:result is None 2026-03-09T17:43:43.861 INFO:teuthology.orchestra.run.vm01.stdout:60129542161 2026-03-09T17:43:43.861 INFO:teuthology.orchestra.run.vm01.stdout:176093659142 2026-03-09T17:43:43.920 INFO:tasks.cephadm.ceph_manager.ceph:need seq 60129542161 got 60129542161 for osd.1 2026-03-09T17:43:43.920 DEBUG:teuthology.parallel:result is None 2026-03-09T17:43:43.959 INFO:tasks.cephadm.ceph_manager.ceph:need seq 176093659142 got 176093659142 for osd.6 2026-03-09T17:43:43.959 DEBUG:teuthology.parallel:result is None 2026-03-09T17:43:43.959 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-09T17:43:43.959 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph pg dump --format=json 2026-03-09T17:43:44.190 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:44.220 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:44 vm01 ceph-mon[52793]: from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/702698574"}]': finished 2026-03-09T17:43:44.220 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:44 vm01 ceph-mon[52793]: osdmap e58: 8 total, 8 up, 8 in 2026-03-09T17:43:44.221 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3015079604' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T17:43:44.221 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2524800812' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/573666992"}]: dispatch 2026-03-09T17:43:44.221 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:44 vm01 ceph-mon[52793]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/573666992"}]: dispatch 2026-03-09T17:43:44.221 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1624713635' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T17:43:44.221 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/4006595158' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T17:43:44.221 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3959248227' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-09T17:43:44.221 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Successfully removed blocklist entry 2026-03-09T17:43:44.221 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Removing blocklisted entry for this host : 192.168.123.101:0/1188799241 2026-03-09T17:43:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:44 vm05 ceph-mon[53831]: from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/702698574"}]': finished 2026-03-09T17:43:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:44 vm05 ceph-mon[53831]: osdmap e58: 8 total, 8 up, 8 in 2026-03-09T17:43:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3015079604' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T17:43:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2524800812' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/573666992"}]: dispatch 2026-03-09T17:43:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:44 vm05 ceph-mon[53831]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/573666992"}]: dispatch 2026-03-09T17:43:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1624713635' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T17:43:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/4006595158' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T17:43:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3959248227' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-09T17:43:44.423 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:43:44.424 INFO:teuthology.orchestra.run.vm01.stderr:dumped all 2026-03-09T17:43:44.493 INFO:teuthology.orchestra.run.vm01.stdout:{"pg_ready":true,"pg_map":{"version":115,"stamp":"2026-03-09T17:43:44.043205+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459688,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":64,"num_read_kb":49,"num_write":63,"num_write_kb":587,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":41,"ondisk_log_size":41,"up":12,"acting":12,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":12,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":6,"kb":167739392,"kb_used":220336,"kb_used_data":5596,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167519056,"statfs":{"total":171765137408,"available":171539513344,"internally_reserved":0,"allocated":5730304,"data_stored":3190761,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12711,"internal_metadata":219663961},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":3,"num_read_kb":3,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.203850"},"pg_stats":[{"pgid":"2.2","version":"50'2","reported_seq":30,"reported_epoch":55,"state":"active+clean","last_fresh":"2026-03-09T17:43:39.841081+0000","last_change":"2026-03-09T17:43:32.938597+0000","last_active":"2026-03-09T17:43:39.841081+0000","last_peered":"2026-03-09T17:43:39.841081+0000","last_clean":"2026-03-09T17:43:39.841081+0000","last_became_active":"2026-03-09T17:43:31.508150+0000","last_became_peered":"2026-03-09T17:43:31.508150+0000","last_unstale":"2026-03-09T17:43:39.841081+0000","last_undegraded":"2026-03-09T17:43:39.841081+0000","last_fullsized":"2026-03-09T17:43:39.841081+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T17:43:29.901234+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T17:43:29.901234+0000","last_clean_scrub_stamp":"2026-03-09T17:43:29.901234+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-10T21:11:43.011348+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00022931699999999999,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,2],"acting":[3,1,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"2.1","version":"48'1","reported_seq":28,"reported_epoch":54,"state":"active+clean","last_fresh":"2026-03-09T17:43:38.994744+0000","last_change":"2026-03-09T17:43:32.940995+0000","last_active":"2026-03-09T17:43:38.994744+0000","last_peered":"2026-03-09T17:43:38.994744+0000","last_clean":"2026-03-09T17:43:38.994744+0000","last_became_active":"2026-03-09T17:43:31.508052+0000","last_became_peered":"2026-03-09T17:43:31.508052+0000","last_unstale":"2026-03-09T17:43:38.994744+0000","last_undegraded":"2026-03-09T17:43:38.994744+0000","last_fullsized":"2026-03-09T17:43:38.994744+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T17:43:29.901234+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T17:43:29.901234+0000","last_clean_scrub_stamp":"2026-03-09T17:43:29.901234+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T02:14:20.096597+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00020661599999999999,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,1,0],"acting":[2,1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.0","version":"50'6","reported_seq":50,"reported_epoch":55,"state":"active+clean","last_fresh":"2026-03-09T17:43:39.841170+0000","last_change":"2026-03-09T17:43:32.938572+0000","last_active":"2026-03-09T17:43:39.841170+0000","last_peered":"2026-03-09T17:43:39.841170+0000","last_clean":"2026-03-09T17:43:39.841170+0000","last_became_active":"2026-03-09T17:43:31.470327+0000","last_became_peered":"2026-03-09T17:43:31.470327+0000","last_unstale":"2026-03-09T17:43:39.841170+0000","last_undegraded":"2026-03-09T17:43:39.841170+0000","last_fullsized":"2026-03-09T17:43:39.841170+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T17:43:29.901234+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T17:43:29.901234+0000","last_clean_scrub_stamp":"2026-03-09T17:43:29.901234+0000","objects_scrubbed":0,"log_size":6,"log_dups_size":0,"ondisk_log_size":6,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-10T21:33:04.475731+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00020549300000000001,"stat_sum":{"num_bytes":389,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":18,"num_read_kb":12,"num_write":4,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,6],"acting":[3,5,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"1.0","version":"22'32","reported_seq":77,"reported_epoch":55,"state":"active+clean","last_fresh":"2026-03-09T17:43:39.841128+0000","last_change":"2026-03-09T17:43:06.849874+0000","last_active":"2026-03-09T17:43:39.841128+0000","last_peered":"2026-03-09T17:43:39.841128+0000","last_clean":"2026-03-09T17:43:39.841128+0000","last_became_active":"2026-03-09T17:43:06.844442+0000","last_became_peered":"2026-03-09T17:43:06.844442+0000","last_unstale":"2026-03-09T17:43:39.841128+0000","last_undegraded":"2026-03-09T17:43:39.841128+0000","last_fullsized":"2026-03-09T17:43:39.841128+0000","mapping_epoch":36,"log_start":"0'0","ondisk_log_start":"0'0","created":20,"last_epoch_clean":37,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T17:42:35.568597+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T17:42:35.568597+0000","last_clean_scrub_stamp":"2026-03-09T17:42:35.568597+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-10T23:51:04.869687+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,2],"acting":[3,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]}],"pool_stats":[{"poolid":2,"num_pg":3,"stat_sum":{"num_bytes":408,"num_objects":3,"num_object_clones":0,"num_object_copies":9,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":3,"num_whiteouts":0,"num_read":18,"num_read_kb":12,"num_write":6,"num_write_kb":3,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":24576,"data_stored":1224,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":9,"ondisk_log_size":9,"up":9,"acting":9,"num_store_stats":6},{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":2314240,"data_stored":2296400,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":5}],"osd_stats":[{"osd":7,"up_from":45,"seq":193273528325,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27392,"kb_used_data":548,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940032,"statfs":{"total":21470642176,"available":21442592768,"internally_reserved":0,"allocated":561152,"data_stored":236531,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1585,"internal_metadata":27457999},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":41,"seq":176093659143,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27380,"kb_used_data":540,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940044,"statfs":{"total":21470642176,"available":21442605056,"internally_reserved":0,"allocated":552960,"data_stored":231218,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":35,"seq":150323855369,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27828,"kb_used_data":980,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939596,"statfs":{"total":21470642176,"available":21442146304,"internally_reserved":0,"allocated":1003520,"data_stored":684697,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":31,"seq":133143986187,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27356,"kb_used_data":512,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940068,"statfs":{"total":21470642176,"available":21442629632,"internally_reserved":0,"allocated":524288,"data_stored":219212,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":26,"seq":111669149709,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27836,"kb_used_data":996,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939588,"statfs":{"total":21470642176,"available":21442138112,"internally_reserved":0,"allocated":1019904,"data_stored":690517,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":19,"seq":81604378639,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27820,"kb_used_data":980,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939604,"statfs":{"total":21470642176,"available":21442154496,"internally_reserved":0,"allocated":1003520,"data_stored":684327,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":14,"seq":60129542161,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27356,"kb_used_data":516,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940068,"statfs":{"total":21470642176,"available":21442629632,"internally_reserved":0,"allocated":528384,"data_stored":219231,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":9,"seq":38654705683,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27368,"kb_used_data":524,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940056,"statfs":{"total":21470642176,"available":21442617344,"internally_reserved":0,"allocated":536576,"data_stored":225028,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":408,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T17:43:44.494 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph pg dump --format=json 2026-03-09T17:43:44.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:44 vm04 ceph-mon[53158]: from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/702698574"}]': finished 2026-03-09T17:43:44.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:44 vm04 ceph-mon[53158]: osdmap e58: 8 total, 8 up, 8 in 2026-03-09T17:43:44.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3015079604' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T17:43:44.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2524800812' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/573666992"}]: dispatch 2026-03-09T17:43:44.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:44 vm04 ceph-mon[53158]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/573666992"}]: dispatch 2026-03-09T17:43:44.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1624713635' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T17:43:44.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/4006595158' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T17:43:44.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3959248227' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-09T17:43:44.668 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:44.723 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:43:44.888 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:43:44.888 INFO:teuthology.orchestra.run.vm01.stderr:dumped all 2026-03-09T17:43:44.941 INFO:teuthology.orchestra.run.vm01.stdout:{"pg_ready":true,"pg_map":{"version":115,"stamp":"2026-03-09T17:43:44.043205+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459688,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":64,"num_read_kb":49,"num_write":63,"num_write_kb":587,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":41,"ondisk_log_size":41,"up":12,"acting":12,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":12,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":6,"kb":167739392,"kb_used":220336,"kb_used_data":5596,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167519056,"statfs":{"total":171765137408,"available":171539513344,"internally_reserved":0,"allocated":5730304,"data_stored":3190761,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12711,"internal_metadata":219663961},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":3,"num_read_kb":3,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.203850"},"pg_stats":[{"pgid":"2.2","version":"50'2","reported_seq":30,"reported_epoch":55,"state":"active+clean","last_fresh":"2026-03-09T17:43:39.841081+0000","last_change":"2026-03-09T17:43:32.938597+0000","last_active":"2026-03-09T17:43:39.841081+0000","last_peered":"2026-03-09T17:43:39.841081+0000","last_clean":"2026-03-09T17:43:39.841081+0000","last_became_active":"2026-03-09T17:43:31.508150+0000","last_became_peered":"2026-03-09T17:43:31.508150+0000","last_unstale":"2026-03-09T17:43:39.841081+0000","last_undegraded":"2026-03-09T17:43:39.841081+0000","last_fullsized":"2026-03-09T17:43:39.841081+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T17:43:29.901234+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T17:43:29.901234+0000","last_clean_scrub_stamp":"2026-03-09T17:43:29.901234+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-10T21:11:43.011348+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00022931699999999999,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,2],"acting":[3,1,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"2.1","version":"48'1","reported_seq":28,"reported_epoch":54,"state":"active+clean","last_fresh":"2026-03-09T17:43:38.994744+0000","last_change":"2026-03-09T17:43:32.940995+0000","last_active":"2026-03-09T17:43:38.994744+0000","last_peered":"2026-03-09T17:43:38.994744+0000","last_clean":"2026-03-09T17:43:38.994744+0000","last_became_active":"2026-03-09T17:43:31.508052+0000","last_became_peered":"2026-03-09T17:43:31.508052+0000","last_unstale":"2026-03-09T17:43:38.994744+0000","last_undegraded":"2026-03-09T17:43:38.994744+0000","last_fullsized":"2026-03-09T17:43:38.994744+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T17:43:29.901234+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T17:43:29.901234+0000","last_clean_scrub_stamp":"2026-03-09T17:43:29.901234+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T02:14:20.096597+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00020661599999999999,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,1,0],"acting":[2,1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.0","version":"50'6","reported_seq":50,"reported_epoch":55,"state":"active+clean","last_fresh":"2026-03-09T17:43:39.841170+0000","last_change":"2026-03-09T17:43:32.938572+0000","last_active":"2026-03-09T17:43:39.841170+0000","last_peered":"2026-03-09T17:43:39.841170+0000","last_clean":"2026-03-09T17:43:39.841170+0000","last_became_active":"2026-03-09T17:43:31.470327+0000","last_became_peered":"2026-03-09T17:43:31.470327+0000","last_unstale":"2026-03-09T17:43:39.841170+0000","last_undegraded":"2026-03-09T17:43:39.841170+0000","last_fullsized":"2026-03-09T17:43:39.841170+0000","mapping_epoch":47,"log_start":"0'0","ondisk_log_start":"0'0","created":47,"last_epoch_clean":48,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T17:43:29.901234+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T17:43:29.901234+0000","last_clean_scrub_stamp":"2026-03-09T17:43:29.901234+0000","objects_scrubbed":0,"log_size":6,"log_dups_size":0,"ondisk_log_size":6,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-10T21:33:04.475731+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00020549300000000001,"stat_sum":{"num_bytes":389,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":18,"num_read_kb":12,"num_write":4,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,6],"acting":[3,5,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"1.0","version":"22'32","reported_seq":77,"reported_epoch":55,"state":"active+clean","last_fresh":"2026-03-09T17:43:39.841128+0000","last_change":"2026-03-09T17:43:06.849874+0000","last_active":"2026-03-09T17:43:39.841128+0000","last_peered":"2026-03-09T17:43:39.841128+0000","last_clean":"2026-03-09T17:43:39.841128+0000","last_became_active":"2026-03-09T17:43:06.844442+0000","last_became_peered":"2026-03-09T17:43:06.844442+0000","last_unstale":"2026-03-09T17:43:39.841128+0000","last_undegraded":"2026-03-09T17:43:39.841128+0000","last_fullsized":"2026-03-09T17:43:39.841128+0000","mapping_epoch":36,"log_start":"0'0","ondisk_log_start":"0'0","created":20,"last_epoch_clean":37,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T17:42:35.568597+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T17:42:35.568597+0000","last_clean_scrub_stamp":"2026-03-09T17:42:35.568597+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-10T23:51:04.869687+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,2],"acting":[3,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]}],"pool_stats":[{"poolid":2,"num_pg":3,"stat_sum":{"num_bytes":408,"num_objects":3,"num_object_clones":0,"num_object_copies":9,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":3,"num_whiteouts":0,"num_read":18,"num_read_kb":12,"num_write":6,"num_write_kb":3,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":24576,"data_stored":1224,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":9,"ondisk_log_size":9,"up":9,"acting":9,"num_store_stats":6},{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":2314240,"data_stored":2296400,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":5}],"osd_stats":[{"osd":7,"up_from":45,"seq":193273528325,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27392,"kb_used_data":548,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940032,"statfs":{"total":21470642176,"available":21442592768,"internally_reserved":0,"allocated":561152,"data_stored":236531,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1585,"internal_metadata":27457999},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":41,"seq":176093659143,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27380,"kb_used_data":540,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940044,"statfs":{"total":21470642176,"available":21442605056,"internally_reserved":0,"allocated":552960,"data_stored":231218,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":35,"seq":150323855369,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27828,"kb_used_data":980,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939596,"statfs":{"total":21470642176,"available":21442146304,"internally_reserved":0,"allocated":1003520,"data_stored":684697,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":31,"seq":133143986187,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27356,"kb_used_data":512,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940068,"statfs":{"total":21470642176,"available":21442629632,"internally_reserved":0,"allocated":524288,"data_stored":219212,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":26,"seq":111669149709,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27836,"kb_used_data":996,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939588,"statfs":{"total":21470642176,"available":21442138112,"internally_reserved":0,"allocated":1019904,"data_stored":690517,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":19,"seq":81604378639,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27820,"kb_used_data":980,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939604,"statfs":{"total":21470642176,"available":21442154496,"internally_reserved":0,"allocated":1003520,"data_stored":684327,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":14,"seq":60129542161,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27356,"kb_used_data":516,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940068,"statfs":{"total":21470642176,"available":21442629632,"internally_reserved":0,"allocated":528384,"data_stored":219231,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":9,"seq":38654705683,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27368,"kb_used_data":524,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940056,"statfs":{"total":21470642176,"available":21442617344,"internally_reserved":0,"allocated":536576,"data_stored":225028,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":408,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T17:43:44.941 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-09T17:43:44.941 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-09T17:43:44.941 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-09T17:43:44.941 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph health --format=json 2026-03-09T17:43:45.115 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:45 vm05 ceph-mon[53831]: pgmap v114: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 767 B/s rd, 0 op/s 2026-03-09T17:43:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:45 vm05 ceph-mon[53831]: from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/573666992"}]': finished 2026-03-09T17:43:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:45 vm05 ceph-mon[53831]: osdmap e59: 8 total, 8 up, 8 in 2026-03-09T17:43:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:45 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2895384165' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1188799241"}]: dispatch 2026-03-09T17:43:45.359 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-mon[52793]: pgmap v114: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 767 B/s rd, 0 op/s 2026-03-09T17:43:45.359 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-mon[52793]: from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/573666992"}]': finished 2026-03-09T17:43:45.359 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-mon[52793]: osdmap e59: 8 total, 8 up, 8 in 2026-03-09T17:43:45.359 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2895384165' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1188799241"}]: dispatch 2026-03-09T17:43:45.359 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Successfully removed blocklist entry 2026-03-09T17:43:45.359 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: * Serving Flask app 'rbd-target-api' (lazy loading) 2026-03-09T17:43:45.359 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: * Environment: production 2026-03-09T17:43:45.359 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-09T17:43:45.359 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: Use a production WSGI server instead. 2026-03-09T17:43:45.359 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: * Debug mode: off 2026-03-09T17:43:45.359 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Reading the configuration object to update local LIO configuration 2026-03-09T17:43:45.359 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Configuration does not have an entry for this host(vm01.local) - nothing to define to LIO 2026-03-09T17:43:45.359 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug * Running on all addresses. 2026-03-09T17:43:45.360 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-09T17:43:45.360 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: * Running on all addresses. 2026-03-09T17:43:45.360 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-09T17:43:45.360 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug * Running on http://[::1]:5000/ (Press CTRL+C to quit) 2026-03-09T17:43:45.360 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: * Running on http://[::1]:5000/ (Press CTRL+C to quit) 2026-03-09T17:43:45.360 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:43:45.360 INFO:teuthology.orchestra.run.vm01.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-09T17:43:45.433 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-09T17:43:45.433 INFO:tasks.cephadm:Setup complete, yielding 2026-03-09T17:43:45.433 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-09T17:43:45.435 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm01.local 2026-03-09T17:43:45.435 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- bash -c 'ceph orch status' 2026-03-09T17:43:45.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:45 vm04 ceph-mon[53158]: pgmap v114: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 767 B/s rd, 0 op/s 2026-03-09T17:43:45.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:45 vm04 ceph-mon[53158]: from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/573666992"}]': finished 2026-03-09T17:43:45.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:45 vm04 ceph-mon[53158]: osdmap e59: 8 total, 8 up, 8 in 2026-03-09T17:43:45.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:45 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2895384165' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1188799241"}]: dispatch 2026-03-09T17:43:45.621 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:45.851 INFO:teuthology.orchestra.run.vm01.stdout:Backend: cephadm 2026-03-09T17:43:45.851 INFO:teuthology.orchestra.run.vm01.stdout:Available: Yes 2026-03-09T17:43:45.851 INFO:teuthology.orchestra.run.vm01.stdout:Paused: No 2026-03-09T17:43:45.906 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- bash -c 'ceph orch ps' 2026-03-09T17:43:46.058 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:43:45 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:43:46.089 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:46.156 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:46 vm01 ceph-mon[52793]: from='client.14664 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:43:46.156 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:46 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:43:46.156 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:46 vm01 ceph-mon[52793]: from='client.24533 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:43:46.156 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:46 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2895384165' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1188799241"}]': finished 2026-03-09T17:43:46.156 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:46 vm01 ceph-mon[52793]: osdmap e60: 8 total, 8 up, 8 in 2026-03-09T17:43:46.156 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:46 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3625602171' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T17:43:46.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:46 vm05 ceph-mon[53831]: from='client.14664 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:43:46.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:46 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:43:46.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:46 vm05 ceph-mon[53831]: from='client.24533 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:43:46.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:46 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2895384165' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1188799241"}]': finished 2026-03-09T17:43:46.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:46 vm05 ceph-mon[53831]: osdmap e60: 8 total, 8 up, 8 in 2026-03-09T17:43:46.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:46 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3625602171' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T17:43:46.322 INFO:teuthology.orchestra.run.vm01.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-09T17:43:46.323 INFO:teuthology.orchestra.run.vm01.stdout:iscsi.iscsi.a vm01 *:5000 running (12s) 10s ago 11s - - 3.9 654f31e6858e 966bf2fe12a9 2026-03-09T17:43:46.323 INFO:teuthology.orchestra.run.vm01.stdout:iscsi.iscsi.b vm05 *:5000 running (11s) 9s ago 10s - - 3.9 654f31e6858e 49ceb8ca0435 2026-03-09T17:43:46.323 INFO:teuthology.orchestra.run.vm01.stdout:mgr.x vm01 *:9283,8765 running (2m) 10s ago 2m - - 19.2.3-678-ge911bdeb 654f31e6858e d63e78dabf11 2026-03-09T17:43:46.323 INFO:teuthology.orchestra.run.vm01.stdout:mon.a vm01 running (2m) 10s ago 2m - 2048M 19.2.3-678-ge911bdeb 654f31e6858e 98555f31d8da 2026-03-09T17:43:46.323 INFO:teuthology.orchestra.run.vm01.stdout:mon.b vm04 running (118s) 55s ago 118s 39.8M 2048M 19.2.3-678-ge911bdeb 654f31e6858e f68aca850be2 2026-03-09T17:43:46.323 INFO:teuthology.orchestra.run.vm01.stdout:mon.c vm05 running (2m) 9s ago 2m - 2048M 19.2.3-678-ge911bdeb 654f31e6858e f158eba845cf 2026-03-09T17:43:46.323 INFO:teuthology.orchestra.run.vm01.stdout:osd.0 vm01 running (97s) 10s ago 97s - 4096M 19.2.3-678-ge911bdeb 654f31e6858e f17d6a09c338 2026-03-09T17:43:46.323 INFO:teuthology.orchestra.run.vm01.stdout:osd.1 vm01 running (87s) 10s ago 87s - 4096M 19.2.3-678-ge911bdeb 654f31e6858e 6b97853eb43b 2026-03-09T17:43:46.323 INFO:teuthology.orchestra.run.vm01.stdout:osd.2 vm04 running (77s) 55s ago 77s 61.2M 1451M 19.2.3-678-ge911bdeb 654f31e6858e e6067fd11c04 2026-03-09T17:43:46.323 INFO:teuthology.orchestra.run.vm01.stdout:osd.3 vm04 running (66s) 55s ago 66s 58.1M 1451M 19.2.3-678-ge911bdeb 654f31e6858e 114082d9a788 2026-03-09T17:43:46.323 INFO:teuthology.orchestra.run.vm01.stdout:osd.4 vm04 running (55s) 55s ago 55s 15.4M 1451M 19.2.3-678-ge911bdeb 654f31e6858e 619ada5a151a 2026-03-09T17:43:46.323 INFO:teuthology.orchestra.run.vm01.stdout:osd.5 vm05 running (46s) 9s ago 46s - 1451M 19.2.3-678-ge911bdeb 654f31e6858e 01007efac186 2026-03-09T17:43:46.323 INFO:teuthology.orchestra.run.vm01.stdout:osd.6 vm05 running (35s) 9s ago 35s - 1451M 19.2.3-678-ge911bdeb 654f31e6858e a7b407416475 2026-03-09T17:43:46.323 INFO:teuthology.orchestra.run.vm01.stdout:osd.7 vm05 running (24s) 9s ago 24s - 1451M 19.2.3-678-ge911bdeb 654f31e6858e 8a1ffa3d0d45 2026-03-09T17:43:46.395 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- bash -c 'ceph orch ls' 2026-03-09T17:43:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:46 vm04 ceph-mon[53158]: from='client.14664 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:43:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:46 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:43:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:46 vm04 ceph-mon[53158]: from='client.24533 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:43:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:46 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2895384165' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1188799241"}]': finished 2026-03-09T17:43:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:46 vm04 ceph-mon[53158]: osdmap e60: 8 total, 8 up, 8 in 2026-03-09T17:43:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:46 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3625602171' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T17:43:46.567 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:46.816 INFO:teuthology.orchestra.run.vm01.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-09T17:43:46.816 INFO:teuthology.orchestra.run.vm01.stdout:iscsi.datapool ?:5000 2/2 10s ago 13s vm01=iscsi.a;vm05=iscsi.b;count:2 2026-03-09T17:43:46.816 INFO:teuthology.orchestra.run.vm01.stdout:mgr 1/1 10s ago 108s vm01=x;count:1 2026-03-09T17:43:46.816 INFO:teuthology.orchestra.run.vm01.stdout:mon 3/3 55s ago 2m vm01:192.168.123.101=a;vm04:192.168.123.104=b;vm05:192.168.123.105=c;count:3 2026-03-09T17:43:46.816 INFO:teuthology.orchestra.run.vm01.stdout:osd 8 55s ago - 2026-03-09T17:43:46.882 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- bash -c 'ceph orch host ls' 2026-03-09T17:43:47.052 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:47.282 INFO:teuthology.orchestra.run.vm01.stdout:HOST ADDR LABELS STATUS 2026-03-09T17:43:47.282 INFO:teuthology.orchestra.run.vm01.stdout:vm01 192.168.123.101 2026-03-09T17:43:47.282 INFO:teuthology.orchestra.run.vm01.stdout:vm04 192.168.123.104 2026-03-09T17:43:47.282 INFO:teuthology.orchestra.run.vm01.stdout:vm05 192.168.123.105 2026-03-09T17:43:47.282 INFO:teuthology.orchestra.run.vm01.stdout:3 hosts in cluster 2026-03-09T17:43:47.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:47 vm05 ceph-mon[53831]: pgmap v117: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:47.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:47 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:43:47.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:47 vm05 ceph-mon[53831]: from='client.14682 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:47.328 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:47 vm01 ceph-mon[52793]: pgmap v117: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:47.328 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:47 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:43:47.328 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:47 vm01 ceph-mon[52793]: from='client.14682 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:47.336 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- bash -c 'ceph orch device ls' 2026-03-09T17:43:47.516 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:43:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:47 vm04 ceph-mon[53158]: pgmap v117: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:43:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:47 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:43:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:47 vm04 ceph-mon[53158]: from='client.14682 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:47.770 INFO:teuthology.orchestra.run.vm01.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-09T17:43:47.770 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 86s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-09T17:43:47.770 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdb hdd DWNBRSTVMM01001 20.0G Yes 86s ago 2026-03-09T17:43:47.770 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdc hdd DWNBRSTVMM01002 20.0G Yes 86s ago 2026-03-09T17:43:47.770 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdd hdd DWNBRSTVMM01003 20.0G No 86s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-09T17:43:47.770 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vde hdd DWNBRSTVMM01004 20.0G No 86s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-09T17:43:47.771 INFO:teuthology.orchestra.run.vm01.stdout:vm04 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 55s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-09T17:43:47.771 INFO:teuthology.orchestra.run.vm01.stdout:vm04 /dev/vdb hdd DWNBRSTVMM04001 20.0G Yes 55s ago 2026-03-09T17:43:47.771 INFO:teuthology.orchestra.run.vm01.stdout:vm04 /dev/vdc hdd DWNBRSTVMM04002 20.0G No 55s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-09T17:43:47.771 INFO:teuthology.orchestra.run.vm01.stdout:vm04 /dev/vdd hdd DWNBRSTVMM04003 20.0G No 55s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-09T17:43:47.771 INFO:teuthology.orchestra.run.vm01.stdout:vm04 /dev/vde hdd DWNBRSTVMM04004 20.0G No 55s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-09T17:43:47.771 INFO:teuthology.orchestra.run.vm01.stdout:vm05 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 23s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-09T17:43:47.771 INFO:teuthology.orchestra.run.vm01.stdout:vm05 /dev/vdb hdd DWNBRSTVMM05001 20.0G Yes 23s ago 2026-03-09T17:43:47.771 INFO:teuthology.orchestra.run.vm01.stdout:vm05 /dev/vdc hdd DWNBRSTVMM05002 20.0G No 23s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-09T17:43:47.771 INFO:teuthology.orchestra.run.vm01.stdout:vm05 /dev/vdd hdd DWNBRSTVMM05003 20.0G No 23s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-09T17:43:47.771 INFO:teuthology.orchestra.run.vm01.stdout:vm05 /dev/vde hdd DWNBRSTVMM05004 20.0G No 23s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-09T17:43:47.824 INFO:teuthology.run_tasks:Running task install... 2026-03-09T17:43:47.826 DEBUG:teuthology.task.install:project ceph 2026-03-09T17:43:47.826 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'}, 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-09T17:43:47.826 DEBUG:teuthology.task.install:config {'extra_system_packages': {'deb': ['open-iscsi', 'multipath-tools', 'python3-xmltodict', 'python3-jmespath'], 'rpm': ['iscsi-initiator-utils', 'device-mapper-multipath', 'bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-09T17:43:47.826 INFO:teuthology.task.install:Using flavor: default 2026-03-09T17:43:47.828 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-09T17:43:47.828 INFO:teuthology.task.install:extra packages: [] 2026-03-09T17:43:47.828 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['open-iscsi', 'multipath-tools', 'python3-xmltodict', 'python3-jmespath'], 'rpm': ['iscsi-initiator-utils', 'device-mapper-multipath', 'bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T17:43:47.829 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T17:43:47.829 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['open-iscsi', 'multipath-tools', 'python3-xmltodict', 'python3-jmespath'], 'rpm': ['iscsi-initiator-utils', 'device-mapper-multipath', 'bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T17:43:47.829 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T17:43:47.829 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['open-iscsi', 'multipath-tools', 'python3-xmltodict', 'python3-jmespath'], 'rpm': ['iscsi-initiator-utils', 'device-mapper-multipath', 'bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T17:43:47.829 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T17:43:48.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:48 vm05 ceph-mon[53831]: from='client.14688 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:48.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:48 vm05 ceph-mon[53831]: from='client.14694 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:48.445 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T17:43:48.445 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T17:43:48.507 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:48 vm01 ceph-mon[52793]: from='client.14688 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:48.507 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:48 vm01 ceph-mon[52793]: from='client.14694 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:48.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:48 vm04 ceph-mon[53158]: from='client.14688 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:48.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:48 vm04 ceph-mon[53158]: from='client.14694 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:48.546 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T17:43:48.546 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T17:43:48.608 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T17:43:48.608 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T17:43:49.034 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T17:43:49.034 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:43:49.034 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T17:43:49.058 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, iscsi-initiator-utils, device-mapper-multipath, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T17:43:49.058 DEBUG:teuthology.orchestra.run.vm01:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T17:43:49.096 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T17:43:49.096 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:43:49.096 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T17:43:49.128 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, iscsi-initiator-utils, device-mapper-multipath, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T17:43:49.129 DEBUG:teuthology.orchestra.run.vm04:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T17:43:49.129 DEBUG:teuthology.orchestra.run.vm01:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T17:43:49.154 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T17:43:49.154 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-09T17:43:49.154 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T17:43:49.183 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, iscsi-initiator-utils, device-mapper-multipath, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T17:43:49.183 DEBUG:teuthology.orchestra.run.vm05:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T17:43:49.204 DEBUG:teuthology.orchestra.run.vm04:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T17:43:49.214 DEBUG:teuthology.orchestra.run.vm01:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T17:43:49.241 INFO:teuthology.orchestra.run.vm01.stdout:check_obsoletes = 1 2026-03-09T17:43:49.242 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean all 2026-03-09T17:43:49.253 DEBUG:teuthology.orchestra.run.vm05:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T17:43:49.281 DEBUG:teuthology.orchestra.run.vm04:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T17:43:49.306 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:49 vm01 ceph-mon[52793]: from='client.14700 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:49.306 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:49 vm01 ceph-mon[52793]: pgmap v118: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.9 KiB/s rd, 1 op/s 2026-03-09T17:43:49.306 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:49 vm01 ceph-mon[52793]: from='client.24554 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:49.329 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:49 vm05 ceph-mon[53831]: from='client.14700 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:49.329 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:49 vm05 ceph-mon[53831]: pgmap v118: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.9 KiB/s rd, 1 op/s 2026-03-09T17:43:49.329 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:49 vm05 ceph-mon[53831]: from='client.24554 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:49.333 DEBUG:teuthology.orchestra.run.vm05:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T17:43:49.350 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:49 vm04 ceph-mon[53158]: from='client.14700 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:49.350 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:49 vm04 ceph-mon[53158]: pgmap v118: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.9 KiB/s rd, 1 op/s 2026-03-09T17:43:49.350 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:49 vm04 ceph-mon[53158]: from='client.24554 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:43:49.352 INFO:teuthology.orchestra.run.vm04.stdout:check_obsoletes = 1 2026-03-09T17:43:49.354 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean all 2026-03-09T17:43:49.398 INFO:teuthology.orchestra.run.vm05.stdout:check_obsoletes = 1 2026-03-09T17:43:49.400 DEBUG:teuthology.orchestra.run.vm05:> sudo yum clean all 2026-03-09T17:43:49.420 INFO:teuthology.orchestra.run.vm01.stdout:41 files removed 2026-03-09T17:43:49.444 DEBUG:teuthology.orchestra.run.vm01:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd iscsi-initiator-utils device-mapper-multipath bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T17:43:49.540 INFO:teuthology.orchestra.run.vm04.stdout:41 files removed 2026-03-09T17:43:49.564 DEBUG:teuthology.orchestra.run.vm04:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd iscsi-initiator-utils device-mapper-multipath bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T17:43:49.575 INFO:teuthology.orchestra.run.vm05.stdout:41 files removed 2026-03-09T17:43:49.600 DEBUG:teuthology.orchestra.run.vm05:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd iscsi-initiator-utils device-mapper-multipath bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T17:43:50.941 INFO:teuthology.orchestra.run.vm01.stdout:ceph packages for x86_64 65 kB/s | 84 kB 00:01 2026-03-09T17:43:50.960 INFO:teuthology.orchestra.run.vm04.stdout:ceph packages for x86_64 69 kB/s | 84 kB 00:01 2026-03-09T17:43:50.968 INFO:teuthology.orchestra.run.vm05.stdout:ceph packages for x86_64 70 kB/s | 84 kB 00:01 2026-03-09T17:43:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:51 vm01 ceph-mon[52793]: pgmap v119: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-09T17:43:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:51 vm04 ceph-mon[53158]: pgmap v119: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-09T17:43:51.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:51 vm05 ceph-mon[53831]: pgmap v119: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-09T17:43:51.926 INFO:teuthology.orchestra.run.vm04.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-09T17:43:51.939 INFO:teuthology.orchestra.run.vm05.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-09T17:43:51.943 INFO:teuthology.orchestra.run.vm01.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-09T17:43:52.904 INFO:teuthology.orchestra.run.vm04.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T17:43:52.904 INFO:teuthology.orchestra.run.vm05.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T17:43:52.905 INFO:teuthology.orchestra.run.vm01.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T17:43:53.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:53 vm01 ceph-mon[52793]: pgmap v120: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:43:53.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:53 vm04 ceph-mon[53158]: pgmap v120: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:43:53.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:53 vm05 ceph-mon[53831]: pgmap v120: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:43:54.527 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - BaseOS 5.6 MB/s | 8.9 MB 00:01 2026-03-09T17:43:54.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:43:54 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:43:55.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:55 vm01 ceph-mon[52793]: pgmap v121: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.1 KiB/s rd, 2 op/s 2026-03-09T17:43:55.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:55 vm04 ceph-mon[53158]: pgmap v121: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.1 KiB/s rd, 2 op/s 2026-03-09T17:43:55.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:55 vm05 ceph-mon[53831]: pgmap v121: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.1 KiB/s rd, 2 op/s 2026-03-09T17:43:56.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:43:55 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:43:56.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:56 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:43:56.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:56 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:43:56.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:56 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:43:57.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:57 vm01 ceph-mon[52793]: pgmap v122: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.9 KiB/s rd, 1 op/s 2026-03-09T17:43:57.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:57 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:43:57.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:57 vm04 ceph-mon[53158]: pgmap v122: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.9 KiB/s rd, 1 op/s 2026-03-09T17:43:57.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:57 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:43:57.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:57 vm05 ceph-mon[53831]: pgmap v122: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.9 KiB/s rd, 1 op/s 2026-03-09T17:43:57.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:57 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:43:59.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:43:59 vm01 ceph-mon[52793]: pgmap v123: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:43:59.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:43:59 vm04 ceph-mon[53158]: pgmap v123: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:43:59.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:43:59 vm05 ceph-mon[53831]: pgmap v123: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:00.518 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - AppStream 5.1 MB/s | 27 MB 00:05 2026-03-09T17:44:01.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:01 vm01 ceph-mon[52793]: pgmap v124: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:01.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:01 vm04 ceph-mon[53158]: pgmap v124: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:01.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:01 vm05 ceph-mon[53831]: pgmap v124: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:03 vm01 ceph-mon[52793]: pgmap v125: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-09T17:44:03.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:03 vm04 ceph-mon[53158]: pgmap v125: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-09T17:44:03.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:03 vm05 ceph-mon[53831]: pgmap v125: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-09T17:44:03.954 INFO:teuthology.orchestra.run.vm05.stdout:CentOS Stream 9 - BaseOS 827 kB/s | 8.9 MB 00:11 2026-03-09T17:44:04.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:44:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:44:05.109 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - BaseOS 748 kB/s | 8.9 MB 00:12 2026-03-09T17:44:05.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:05 vm01 ceph-mon[52793]: pgmap v126: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-09T17:44:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:05 vm04 ceph-mon[53158]: pgmap v126: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-09T17:44:05.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:05 vm05 ceph-mon[53831]: pgmap v126: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-09T17:44:06.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:44:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:44:06.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:06 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:06.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:06 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:06 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:06.926 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - CRB 2.2 MB/s | 8.0 MB 00:03 2026-03-09T17:44:07.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:07 vm01 ceph-mon[52793]: pgmap v127: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-09T17:44:07.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:07 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:07.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:07 vm04 ceph-mon[53158]: pgmap v127: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-09T17:44:07.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:07 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:07.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:07 vm05 ceph-mon[53831]: pgmap v127: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-09T17:44:07.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:07 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:08.156 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - Extras packages 55 kB/s | 20 kB 00:00 2026-03-09T17:44:08.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:08 vm01 ceph-mon[52793]: pgmap v128: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-09T17:44:08.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:08 vm04 ceph-mon[53158]: pgmap v128: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-09T17:44:08.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:08 vm05 ceph-mon[53831]: pgmap v128: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-09T17:44:09.040 INFO:teuthology.orchestra.run.vm01.stdout:Extra Packages for Enterprise Linux 25 MB/s | 20 MB 00:00 2026-03-09T17:44:09.742 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - AppStream 6.8 MB/s | 27 MB 00:03 2026-03-09T17:44:10.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:10 vm01 ceph-mon[52793]: pgmap v129: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-09T17:44:10.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:10 vm04 ceph-mon[53158]: pgmap v129: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-09T17:44:10.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:10 vm05 ceph-mon[53831]: pgmap v129: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-09T17:44:11.691 INFO:teuthology.orchestra.run.vm05.stdout:CentOS Stream 9 - AppStream 3.8 MB/s | 27 MB 00:07 2026-03-09T17:44:12.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:12 vm01 ceph-mon[52793]: pgmap v130: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-09T17:44:12.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:12 vm04 ceph-mon[53158]: pgmap v130: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-09T17:44:12.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:12 vm05 ceph-mon[53831]: pgmap v130: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-09T17:44:14.485 INFO:teuthology.orchestra.run.vm01.stdout:lab-extras 65 kB/s | 50 kB 00:00 2026-03-09T17:44:14.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:44:14 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:44:14.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:14 vm01 ceph-mon[52793]: pgmap v131: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:14 vm04 ceph-mon[53158]: pgmap v131: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:14.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:14 vm05 ceph-mon[53831]: pgmap v131: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:15.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:15 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:15.789 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:15 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:15.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:15 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:15.909 INFO:teuthology.orchestra.run.vm01.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T17:44:15.909 INFO:teuthology.orchestra.run.vm01.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T17:44:15.914 INFO:teuthology.orchestra.run.vm01.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T17:44:15.915 INFO:teuthology.orchestra.run.vm01.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T17:44:15.945 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:44:15.950 INFO:teuthology.orchestra.run.vm01.stdout:======================================================================================= 2026-03-09T17:44:15.950 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T17:44:15.950 INFO:teuthology.orchestra.run.vm01.stdout:======================================================================================= 2026-03-09T17:44:15.950 INFO:teuthology.orchestra.run.vm01.stdout:Installing: 2026-03-09T17:44:15.950 INFO:teuthology.orchestra.run.vm01.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T17:44:15.950 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T17:44:15.950 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T17:44:15.950 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T17:44:15.950 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: device-mapper-multipath x86_64 0.8.7-45.el9 baseos 156 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: iscsi-initiator-utils x86_64 6.2.1.11-0.git4b3e853.el9 baseos 392 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout:Upgrading: 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: kpartx x86_64 0.8.7-45.el9 baseos 49 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout:Installing dependencies: 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: device-mapper-multipath-libs x86_64 0.8.7-45.el9 baseos 289 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: iscsi-initiator-utils-iscsiuio x86_64 6.2.1.11-0.git4b3e853.el9 baseos 81 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: isns-utils-libs x86_64 0.101-4.el9 baseos 101 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T17:44:15.951 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T17:44:15.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout:Installing weak dependencies: 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout:======================================================================================= 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout:Install 140 Packages 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout:Upgrade 3 Packages 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:15.953 INFO:teuthology.orchestra.run.vm01.stdout:Total download size: 212 M 2026-03-09T17:44:15.954 INFO:teuthology.orchestra.run.vm01.stdout:Downloading Packages: 2026-03-09T17:44:16.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:44:15 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:44:16.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:16 vm01 ceph-mon[52793]: pgmap v132: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:16.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:16 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:16.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:16 vm04 ceph-mon[53158]: pgmap v132: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:16.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:16 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:16.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:16 vm05 ceph-mon[53831]: pgmap v132: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:16.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:16 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:17.363 INFO:teuthology.orchestra.run.vm01.stdout:(1/143): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 13 kB/s | 6.5 kB 00:00 2026-03-09T17:44:18.136 INFO:teuthology.orchestra.run.vm01.stdout:(2/143): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.5 MB/s | 1.2 MB 00:00 2026-03-09T17:44:18.250 INFO:teuthology.orchestra.run.vm01.stdout:(3/143): ceph-immutable-object-cache-19.2.3-678 1.3 MB/s | 145 kB 00:00 2026-03-09T17:44:18.558 INFO:teuthology.orchestra.run.vm01.stdout:(4/143): ceph-base-19.2.3-678.ge911bdeb.el9.x86 3.2 MB/s | 5.5 MB 00:01 2026-03-09T17:44:18.591 INFO:teuthology.orchestra.run.vm01.stdout:(5/143): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 7.1 MB/s | 2.4 MB 00:00 2026-03-09T17:44:18.704 INFO:teuthology.orchestra.run.vm01.stdout:(6/143): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 7.4 MB/s | 1.1 MB 00:00 2026-03-09T17:44:18.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:18 vm04 ceph-mon[53158]: pgmap v133: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:18.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:18 vm05 ceph-mon[53831]: pgmap v133: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:18.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:18 vm01 ceph-mon[52793]: pgmap v133: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:19.023 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - CRB 1.3 MB/s | 8.0 MB 00:06 2026-03-09T17:44:19.078 INFO:teuthology.orchestra.run.vm01.stdout:(7/143): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 9.7 MB/s | 4.7 MB 00:00 2026-03-09T17:44:19.914 INFO:teuthology.orchestra.run.vm01.stdout:(8/143): ceph-radosgw-19.2.3-678.ge911bdeb.el9. 13 MB/s | 11 MB 00:00 2026-03-09T17:44:20.079 INFO:teuthology.orchestra.run.vm01.stdout:(9/143): ceph-common-19.2.3-678.ge911bdeb.el9.x 6.7 MB/s | 22 MB 00:03 2026-03-09T17:44:20.125 INFO:teuthology.orchestra.run.vm01.stdout:(10/143): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 12 MB/s | 17 MB 00:01 2026-03-09T17:44:20.126 INFO:teuthology.orchestra.run.vm01.stdout:(11/143): ceph-selinux-19.2.3-678.ge911bdeb.el9 118 kB/s | 25 kB 00:00 2026-03-09T17:44:20.296 INFO:teuthology.orchestra.run.vm01.stdout:(12/143): libcephfs-devel-19.2.3-678.ge911bdeb. 198 kB/s | 34 kB 00:00 2026-03-09T17:44:20.337 INFO:teuthology.orchestra.run.vm01.stdout:(13/143): libcephfs2-19.2.3-678.ge911bdeb.el9.x 4.7 MB/s | 1.0 MB 00:00 2026-03-09T17:44:20.411 INFO:teuthology.orchestra.run.vm01.stdout:(14/143): libcephsqlite-19.2.3-678.ge911bdeb.el 1.4 MB/s | 163 kB 00:00 2026-03-09T17:44:20.455 INFO:teuthology.orchestra.run.vm01.stdout:(15/143): librados-devel-19.2.3-678.ge911bdeb.e 1.0 MB/s | 127 kB 00:00 2026-03-09T17:44:20.462 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - Extras packages 35 kB/s | 20 kB 00:00 2026-03-09T17:44:20.528 INFO:teuthology.orchestra.run.vm01.stdout:(16/143): libradosstriper1-19.2.3-678.ge911bdeb 4.2 MB/s | 503 kB 00:00 2026-03-09T17:44:20.640 INFO:teuthology.orchestra.run.vm01.stdout:(17/143): python3-ceph-argparse-19.2.3-678.ge91 403 kB/s | 45 kB 00:00 2026-03-09T17:44:20.754 INFO:teuthology.orchestra.run.vm01.stdout:(18/143): python3-ceph-common-19.2.3-678.ge911b 1.2 MB/s | 142 kB 00:00 2026-03-09T17:44:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:20 vm04 ceph-mon[53158]: pgmap v134: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:20.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:20 vm05 ceph-mon[53831]: pgmap v134: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:20.831 INFO:teuthology.orchestra.run.vm01.stdout:(19/143): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 14 MB/s | 5.4 MB 00:00 2026-03-09T17:44:20.867 INFO:teuthology.orchestra.run.vm01.stdout:(20/143): python3-cephfs-19.2.3-678.ge911bdeb.e 1.4 MB/s | 165 kB 00:00 2026-03-09T17:44:20.951 INFO:teuthology.orchestra.run.vm01.stdout:(21/143): python3-rados-19.2.3-678.ge911bdeb.el 2.6 MB/s | 323 kB 00:00 2026-03-09T17:44:20.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:20 vm01 ceph-mon[52793]: pgmap v134: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:20.982 INFO:teuthology.orchestra.run.vm01.stdout:(22/143): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.6 MB/s | 303 kB 00:00 2026-03-09T17:44:21.070 INFO:teuthology.orchestra.run.vm01.stdout:(23/143): python3-rgw-19.2.3-678.ge911bdeb.el9. 846 kB/s | 100 kB 00:00 2026-03-09T17:44:21.095 INFO:teuthology.orchestra.run.vm01.stdout:(24/143): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 750 kB/s | 85 kB 00:00 2026-03-09T17:44:21.224 INFO:teuthology.orchestra.run.vm01.stdout:(25/143): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.3 MB/s | 171 kB 00:00 2026-03-09T17:44:21.338 INFO:teuthology.orchestra.run.vm01.stdout:(26/143): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 12 MB/s | 3.1 MB 00:00 2026-03-09T17:44:21.339 INFO:teuthology.orchestra.run.vm04.stdout:Extra Packages for Enterprise Linux 26 MB/s | 20 MB 00:00 2026-03-09T17:44:21.339 INFO:teuthology.orchestra.run.vm01.stdout:(27/143): ceph-grafana-dashboards-19.2.3-678.ge 271 kB/s | 31 kB 00:00 2026-03-09T17:44:21.484 INFO:teuthology.orchestra.run.vm01.stdout:(28/143): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.0 MB/s | 150 kB 00:00 2026-03-09T17:44:21.597 INFO:teuthology.orchestra.run.vm01.stdout:(29/143): ceph-mgr-dashboard-19.2.3-678.ge911bd 15 MB/s | 3.8 MB 00:00 2026-03-09T17:44:21.710 INFO:teuthology.orchestra.run.vm01.stdout:(30/143): ceph-mgr-modules-core-19.2.3-678.ge91 2.2 MB/s | 253 kB 00:00 2026-03-09T17:44:21.821 INFO:teuthology.orchestra.run.vm01.stdout:(31/143): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 444 kB/s | 49 kB 00:00 2026-03-09T17:44:21.884 INFO:teuthology.orchestra.run.vm01.stdout:(32/143): ceph-mgr-diskprediction-local-19.2.3- 18 MB/s | 7.4 MB 00:00 2026-03-09T17:44:21.933 INFO:teuthology.orchestra.run.vm01.stdout:(33/143): ceph-prometheus-alerts-19.2.3-678.ge9 151 kB/s | 17 kB 00:00 2026-03-09T17:44:22.004 INFO:teuthology.orchestra.run.vm01.stdout:(34/143): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 299 kB 00:00 2026-03-09T17:44:22.052 INFO:teuthology.orchestra.run.vm01.stdout:(35/143): cephadm-19.2.3-678.ge911bdeb.el9.noar 6.3 MB/s | 769 kB 00:00 2026-03-09T17:44:22.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:22 vm05 ceph-mon[53831]: pgmap v135: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:22.960 INFO:teuthology.orchestra.run.vm01.stdout:(36/143): device-mapper-multipath-0.8.7-45.el9. 172 kB/s | 156 kB 00:00 2026-03-09T17:44:22.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:22 vm01 ceph-mon[52793]: pgmap v135: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:23.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:22 vm04 ceph-mon[53158]: pgmap v135: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:23.041 INFO:teuthology.orchestra.run.vm01.stdout:(37/143): device-mapper-multipath-libs-0.8.7-45 3.5 MB/s | 289 kB 00:00 2026-03-09T17:44:23.127 INFO:teuthology.orchestra.run.vm01.stdout:(38/143): cryptsetup-2.8.1-3.el9.x86_64.rpm 313 kB/s | 351 kB 00:01 2026-03-09T17:44:23.164 INFO:teuthology.orchestra.run.vm01.stdout:(39/143): iscsi-initiator-utils-iscsiuio-6.2.1. 2.1 MB/s | 81 kB 00:00 2026-03-09T17:44:23.204 INFO:teuthology.orchestra.run.vm01.stdout:(40/143): isns-utils-libs-0.101-4.el9.x86_64.rp 2.5 MB/s | 101 kB 00:00 2026-03-09T17:44:23.230 INFO:teuthology.orchestra.run.vm01.stdout:(41/143): iscsi-initiator-utils-6.2.1.11-0.git4 2.0 MB/s | 392 kB 00:00 2026-03-09T17:44:23.406 INFO:teuthology.orchestra.run.vm05.stdout:CentOS Stream 9 - CRB 915 kB/s | 8.0 MB 00:08 2026-03-09T17:44:23.547 INFO:teuthology.orchestra.run.vm01.stdout:(42/143): ledmon-libs-1.1.0-3.el9.x86_64.rpm 118 kB/s | 40 kB 00:00 2026-03-09T17:44:23.632 INFO:teuthology.orchestra.run.vm01.stdout:(43/143): libconfig-1.7.2-9.el9.x86_64.rpm 179 kB/s | 72 kB 00:00 2026-03-09T17:44:24.498 INFO:teuthology.orchestra.run.vm01.stdout:(44/143): libgfortran-11.5.0-14.el9.x86_64.rpm 836 kB/s | 794 kB 00:00 2026-03-09T17:44:24.525 INFO:teuthology.orchestra.run.vm01.stdout:(45/143): mailcap-2.1.49-5.el9.noarch.rpm 1.2 MB/s | 33 kB 00:00 2026-03-09T17:44:24.556 INFO:teuthology.orchestra.run.vm01.stdout:(46/143): libquadmath-11.5.0-14.el9.x86_64.rpm 200 kB/s | 184 kB 00:00 2026-03-09T17:44:24.766 INFO:teuthology.orchestra.run.vm01.stdout:(47/143): python3-cffi-1.14.5-5.el9.x86_64.rpm 1.2 MB/s | 253 kB 00:00 2026-03-09T17:44:24.793 INFO:teuthology.orchestra.run.vm01.stdout:(48/143): pciutils-3.7.0-7.el9.x86_64.rpm 348 kB/s | 93 kB 00:00 2026-03-09T17:44:24.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:24 vm05 ceph-mon[53831]: pgmap v136: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:24.821 INFO:teuthology.orchestra.run.vm01.stdout:(49/143): python3-ply-3.11-14.el9.noarch.rpm 3.7 MB/s | 106 kB 00:00 2026-03-09T17:44:24.834 INFO:teuthology.orchestra.run.vm05.stdout:CentOS Stream 9 - Extras packages 34 kB/s | 20 kB 00:00 2026-03-09T17:44:24.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:44:24 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:44:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:24 vm01 ceph-mon[52793]: pgmap v136: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:25.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:24 vm04 ceph-mon[53158]: pgmap v136: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:25.293 INFO:teuthology.orchestra.run.vm01.stdout:(50/143): python3-cryptography-36.0.1-5.el9.x86 2.4 MB/s | 1.2 MB 00:00 2026-03-09T17:44:25.481 INFO:teuthology.orchestra.run.vm01.stdout:(51/143): python3-pycparser-2.20-6.el9.noarch.r 205 kB/s | 135 kB 00:00 2026-03-09T17:44:25.617 INFO:teuthology.orchestra.run.vm01.stdout:(52/143): ceph-test-19.2.3-678.ge911bdeb.el9.x8 9.0 MB/s | 50 MB 00:05 2026-03-09T17:44:25.772 INFO:teuthology.orchestra.run.vm01.stdout:(53/143): python3-requests-2.25.1-10.el9.noarch 434 kB/s | 126 kB 00:00 2026-03-09T17:44:25.797 INFO:teuthology.orchestra.run.vm05.stdout:Extra Packages for Enterprise Linux 23 MB/s | 20 MB 00:00 2026-03-09T17:44:25.803 INFO:teuthology.orchestra.run.vm01.stdout:(54/143): unzip-6.0-59.el9.x86_64.rpm 5.7 MB/s | 182 kB 00:00 2026-03-09T17:44:25.836 INFO:teuthology.orchestra.run.vm01.stdout:(55/143): zip-3.0-35.el9.x86_64.rpm 8.0 MB/s | 266 kB 00:00 2026-03-09T17:44:25.843 INFO:teuthology.orchestra.run.vm01.stdout:(56/143): python3-pyparsing-2.4.7-9.el9.noarch. 274 kB/s | 150 kB 00:00 2026-03-09T17:44:25.963 INFO:teuthology.orchestra.run.vm01.stdout:(57/143): flexiblas-3.0.4-9.el9.x86_64.rpm 247 kB/s | 30 kB 00:00 2026-03-09T17:44:25.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:25 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:26.015 INFO:teuthology.orchestra.run.vm01.stdout:(58/143): boost-program-options-1.75.0-13.el9.x 582 kB/s | 104 kB 00:00 2026-03-09T17:44:26.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:25 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:26.066 INFO:teuthology.orchestra.run.vm01.stdout:(59/143): flexiblas-openblas-openmp-3.0.4-9.el9 293 kB/s | 15 kB 00:00 2026-03-09T17:44:26.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:25 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:26.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:44:25 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:44:26.206 INFO:teuthology.orchestra.run.vm01.stdout:(60/143): libnbd-1.20.3-4.el9.x86_64.rpm 1.1 MB/s | 164 kB 00:00 2026-03-09T17:44:26.260 INFO:teuthology.orchestra.run.vm01.stdout:(61/143): libpmemobj-1.12.1-1.el9.x86_64.rpm 2.9 MB/s | 160 kB 00:00 2026-03-09T17:44:26.307 INFO:teuthology.orchestra.run.vm01.stdout:(62/143): librabbitmq-0.11.0-7.el9.x86_64.rpm 962 kB/s | 45 kB 00:00 2026-03-09T17:44:26.355 INFO:teuthology.orchestra.run.vm01.stdout:(63/143): python3-urllib3-1.26.5-7.el9.noarch.r 295 kB/s | 218 kB 00:00 2026-03-09T17:44:26.397 INFO:teuthology.orchestra.run.vm04.stdout:lab-extras 65 kB/s | 50 kB 00:00 2026-03-09T17:44:26.481 INFO:teuthology.orchestra.run.vm01.stdout:(64/143): librdkafka-1.6.1-102.el9.x86_64.rpm 3.7 MB/s | 662 kB 00:00 2026-03-09T17:44:26.556 INFO:teuthology.orchestra.run.vm01.stdout:(65/143): flexiblas-netlib-3.0.4-9.el9.x86_64.r 5.0 MB/s | 3.0 MB 00:00 2026-03-09T17:44:26.558 INFO:teuthology.orchestra.run.vm01.stdout:(66/143): libstoragemgmt-1.10.1-1.el9.x86_64.rp 1.2 MB/s | 246 kB 00:00 2026-03-09T17:44:26.559 INFO:teuthology.orchestra.run.vm01.stdout:(67/143): libxslt-1.1.34-12.el9.x86_64.rpm 2.9 MB/s | 233 kB 00:00 2026-03-09T17:44:26.610 INFO:teuthology.orchestra.run.vm01.stdout:(68/143): lttng-ust-2.12.0-6.el9.x86_64.rpm 5.3 MB/s | 292 kB 00:00 2026-03-09T17:44:26.616 INFO:teuthology.orchestra.run.vm01.stdout:(69/143): lua-5.4.4-4.el9.x86_64.rpm 3.2 MB/s | 188 kB 00:00 2026-03-09T17:44:26.623 INFO:teuthology.orchestra.run.vm01.stdout:(70/143): openblas-0.3.29-1.el9.x86_64.rpm 658 kB/s | 42 kB 00:00 2026-03-09T17:44:26.822 INFO:teuthology.orchestra.run.vm01.stdout:(71/143): protobuf-3.14.0-17.el9.x86_64.rpm 4.9 MB/s | 1.0 MB 00:00 2026-03-09T17:44:26.905 INFO:teuthology.orchestra.run.vm01.stdout:(72/143): python3-devel-3.9.25-3.el9.x86_64.rpm 2.9 MB/s | 244 kB 00:00 2026-03-09T17:44:26.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:26 vm01 ceph-mon[52793]: pgmap v137: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:26.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:26 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:27.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:26 vm04 ceph-mon[53158]: pgmap v137: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:27.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:26 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:27.043 INFO:teuthology.orchestra.run.vm01.stdout:(73/143): python3-jinja2-2.11.3-8.el9.noarch.rp 1.8 MB/s | 249 kB 00:00 2026-03-09T17:44:27.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:26 vm05 ceph-mon[53831]: pgmap v137: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:27.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:26 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:27.090 INFO:teuthology.orchestra.run.vm01.stdout:(74/143): python3-jmespath-1.0.1-1.el9.noarch.r 1.0 MB/s | 48 kB 00:00 2026-03-09T17:44:27.195 INFO:teuthology.orchestra.run.vm01.stdout:(75/143): python3-babel-2.9.1-2.el9.noarch.rpm 10 MB/s | 6.0 MB 00:00 2026-03-09T17:44:27.197 INFO:teuthology.orchestra.run.vm01.stdout:(76/143): python3-libstoragemgmt-1.10.1-1.el9.x 1.6 MB/s | 177 kB 00:00 2026-03-09T17:44:27.255 INFO:teuthology.orchestra.run.vm01.stdout:(77/143): python3-markupsafe-1.1.1-12.el9.x86_6 601 kB/s | 35 kB 00:00 2026-03-09T17:44:27.275 INFO:teuthology.orchestra.run.vm01.stdout:(78/143): python3-mako-1.1.4-6.el9.noarch.rpm 2.1 MB/s | 172 kB 00:00 2026-03-09T17:44:27.400 INFO:teuthology.orchestra.run.vm01.stdout:(79/143): openblas-openmp-0.3.29-1.el9.x86_64.r 6.7 MB/s | 5.3 MB 00:00 2026-03-09T17:44:27.407 INFO:teuthology.orchestra.run.vm01.stdout:(80/143): python3-numpy-f2py-1.23.5-2.el9.x86_6 3.3 MB/s | 442 kB 00:00 2026-03-09T17:44:27.450 INFO:teuthology.orchestra.run.vm01.stdout:(81/143): python3-packaging-20.9-5.el9.noarch.r 1.5 MB/s | 77 kB 00:00 2026-03-09T17:44:27.477 INFO:teuthology.orchestra.run.vm01.stdout:(82/143): python3-protobuf-3.14.0-17.el9.noarch 3.7 MB/s | 267 kB 00:00 2026-03-09T17:44:27.535 INFO:teuthology.orchestra.run.vm01.stdout:(83/143): python3-pyasn1-0.4.8-7.el9.noarch.rpm 1.8 MB/s | 157 kB 00:00 2026-03-09T17:44:27.545 INFO:teuthology.orchestra.run.vm01.stdout:(84/143): python3-pyasn1-modules-0.4.8-7.el9.no 4.0 MB/s | 277 kB 00:00 2026-03-09T17:44:27.612 INFO:teuthology.orchestra.run.vm01.stdout:(85/143): python3-requests-oauthlib-1.3.0-12.el 697 kB/s | 54 kB 00:00 2026-03-09T17:44:27.661 INFO:teuthology.orchestra.run.vm01.stdout:(86/143): python3-toml-0.10.2-6.el9.noarch.rpm 849 kB/s | 42 kB 00:00 2026-03-09T17:44:27.752 INFO:teuthology.orchestra.run.vm04.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T17:44:27.753 INFO:teuthology.orchestra.run.vm04.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T17:44:27.759 INFO:teuthology.orchestra.run.vm04.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T17:44:27.759 INFO:teuthology.orchestra.run.vm04.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T17:44:27.789 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout:======================================================================================= 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout:======================================================================================= 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout:Installing: 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: device-mapper-multipath x86_64 0.8.7-45.el9 baseos 156 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: iscsi-initiator-utils x86_64 6.2.1.11-0.git4b3e853.el9 baseos 392 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T17:44:27.794 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout:Upgrading: 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: kpartx x86_64 0.8.7-45.el9 baseos 49 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout:Installing dependencies: 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: device-mapper-multipath-libs x86_64 0.8.7-45.el9 baseos 289 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: iscsi-initiator-utils-iscsiuio x86_64 6.2.1.11-0.git4b3e853.el9 baseos 81 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: isns-utils-libs x86_64 0.101-4.el9 baseos 101 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T17:44:27.795 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T17:44:27.796 INFO:teuthology.orchestra.run.vm04.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T17:44:27.797 INFO:teuthology.orchestra.run.vm04.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T17:44:27.797 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T17:44:27.797 INFO:teuthology.orchestra.run.vm04.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T17:44:27.797 INFO:teuthology.orchestra.run.vm04.stdout:Installing weak dependencies: 2026-03-09T17:44:27.797 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T17:44:27.797 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:44:27.797 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T17:44:27.797 INFO:teuthology.orchestra.run.vm04.stdout:======================================================================================= 2026-03-09T17:44:27.797 INFO:teuthology.orchestra.run.vm04.stdout:Install 140 Packages 2026-03-09T17:44:27.797 INFO:teuthology.orchestra.run.vm04.stdout:Upgrade 3 Packages 2026-03-09T17:44:27.797 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:44:27.797 INFO:teuthology.orchestra.run.vm04.stdout:Total download size: 212 M 2026-03-09T17:44:27.797 INFO:teuthology.orchestra.run.vm04.stdout:Downloading Packages: 2026-03-09T17:44:27.821 INFO:teuthology.orchestra.run.vm01.stdout:(87/143): qatlib-25.08.0-2.el9.x86_64.rpm 1.5 MB/s | 240 kB 00:00 2026-03-09T17:44:27.942 INFO:teuthology.orchestra.run.vm01.stdout:(88/143): qatlib-service-25.08.0-2.el9.x86_64.r 306 kB/s | 37 kB 00:00 2026-03-09T17:44:28.001 INFO:teuthology.orchestra.run.vm01.stdout:(89/143): qatzip-libs-1.3.1-1.el9.x86_64.rpm 1.1 MB/s | 66 kB 00:00 2026-03-09T17:44:28.092 INFO:teuthology.orchestra.run.vm01.stdout:(90/143): socat-1.7.4.1-8.el9.x86_64.rpm 3.2 MB/s | 303 kB 00:00 2026-03-09T17:44:28.141 INFO:teuthology.orchestra.run.vm01.stdout:(91/143): xmlstarlet-1.6.1-20.el9.x86_64.rpm 1.3 MB/s | 64 kB 00:00 2026-03-09T17:44:28.260 INFO:teuthology.orchestra.run.vm01.stdout:(92/143): lua-devel-5.4.4-4.el9.x86_64.rpm 189 kB/s | 22 kB 00:00 2026-03-09T17:44:28.401 INFO:teuthology.orchestra.run.vm01.stdout:(93/143): protobuf-compiler-3.14.0-17.el9.x86_6 6.0 MB/s | 862 kB 00:00 2026-03-09T17:44:28.416 INFO:teuthology.orchestra.run.vm01.stdout:(94/143): abseil-cpp-20211102.0-4.el9.x86_64.rp 38 MB/s | 551 kB 00:00 2026-03-09T17:44:28.424 INFO:teuthology.orchestra.run.vm01.stdout:(95/143): gperftools-libs-2.9.1-3.el9.x86_64.rp 41 MB/s | 308 kB 00:00 2026-03-09T17:44:28.426 INFO:teuthology.orchestra.run.vm01.stdout:(96/143): grpc-data-1.46.7-10.el9.noarch.rpm 7.5 MB/s | 19 kB 00:00 2026-03-09T17:44:28.501 INFO:teuthology.orchestra.run.vm01.stdout:(97/143): libarrow-9.0.0-15.el9.x86_64.rpm 60 MB/s | 4.4 MB 00:00 2026-03-09T17:44:28.503 INFO:teuthology.orchestra.run.vm01.stdout:(98/143): libarrow-doc-9.0.0-15.el9.noarch.rpm 10 MB/s | 25 kB 00:00 2026-03-09T17:44:28.507 INFO:teuthology.orchestra.run.vm01.stdout:(99/143): liboath-2.6.12-1.el9.x86_64.rpm 17 MB/s | 49 kB 00:00 2026-03-09T17:44:28.510 INFO:teuthology.orchestra.run.vm01.stdout:(100/143): libunwind-1.6.2-1.el9.x86_64.rpm 22 MB/s | 67 kB 00:00 2026-03-09T17:44:28.514 INFO:teuthology.orchestra.run.vm01.stdout:(101/143): luarocks-3.9.2-5.el9.noarch.rpm 39 MB/s | 151 kB 00:00 2026-03-09T17:44:28.528 INFO:teuthology.orchestra.run.vm01.stdout:(102/143): parquet-libs-9.0.0-15.el9.x86_64.rpm 60 MB/s | 838 kB 00:00 2026-03-09T17:44:28.537 INFO:teuthology.orchestra.run.vm01.stdout:(103/143): python3-asyncssh-2.13.2-5.el9.noarch 60 MB/s | 548 kB 00:00 2026-03-09T17:44:28.540 INFO:teuthology.orchestra.run.vm01.stdout:(104/143): python3-autocommand-2.2.2-8.el9.noar 11 MB/s | 29 kB 00:00 2026-03-09T17:44:28.543 INFO:teuthology.orchestra.run.vm01.stdout:(105/143): python3-backports-tarfile-1.2.0-1.el 19 MB/s | 60 kB 00:00 2026-03-09T17:44:28.546 INFO:teuthology.orchestra.run.vm01.stdout:(106/143): python3-bcrypt-3.2.2-1.el9.x86_64.rp 16 MB/s | 43 kB 00:00 2026-03-09T17:44:28.549 INFO:teuthology.orchestra.run.vm01.stdout:(107/143): python3-cachetools-4.2.4-1.el9.noarc 13 MB/s | 32 kB 00:00 2026-03-09T17:44:28.551 INFO:teuthology.orchestra.run.vm01.stdout:(108/143): python3-certifi-2023.05.07-4.el9.noa 6.0 MB/s | 14 kB 00:00 2026-03-09T17:44:28.556 INFO:teuthology.orchestra.run.vm01.stdout:(109/143): python3-cheroot-10.0.1-4.el9.noarch. 42 MB/s | 173 kB 00:00 2026-03-09T17:44:28.562 INFO:teuthology.orchestra.run.vm01.stdout:(110/143): python3-cherrypy-18.6.1-2.el9.noarch 58 MB/s | 358 kB 00:00 2026-03-09T17:44:28.568 INFO:teuthology.orchestra.run.vm01.stdout:(111/143): python3-google-auth-2.45.0-1.el9.noa 45 MB/s | 254 kB 00:00 2026-03-09T17:44:28.619 INFO:teuthology.orchestra.run.vm01.stdout:(112/143): python3-grpcio-1.46.7-10.el9.x86_64. 40 MB/s | 2.0 MB 00:00 2026-03-09T17:44:28.624 INFO:teuthology.orchestra.run.vm01.stdout:(113/143): python3-grpcio-tools-1.46.7-10.el9.x 36 MB/s | 144 kB 00:00 2026-03-09T17:44:28.626 INFO:teuthology.orchestra.run.vm01.stdout:(114/143): python3-jaraco-8.2.1-3.el9.noarch.rp 4.3 MB/s | 11 kB 00:00 2026-03-09T17:44:28.629 INFO:teuthology.orchestra.run.vm01.stdout:(115/143): python3-jaraco-classes-3.2.1-5.el9.n 7.2 MB/s | 18 kB 00:00 2026-03-09T17:44:28.632 INFO:teuthology.orchestra.run.vm01.stdout:(116/143): python3-jaraco-collections-3.0.0-8.e 8.8 MB/s | 23 kB 00:00 2026-03-09T17:44:28.635 INFO:teuthology.orchestra.run.vm01.stdout:(117/143): python3-jaraco-context-6.0.1-3.el9.n 8.2 MB/s | 20 kB 00:00 2026-03-09T17:44:28.637 INFO:teuthology.orchestra.run.vm01.stdout:(118/143): python3-jaraco-functools-3.5.0-2.el9 7.7 MB/s | 19 kB 00:00 2026-03-09T17:44:28.640 INFO:teuthology.orchestra.run.vm01.stdout:(119/143): python3-jaraco-text-4.0.0-2.el9.noar 10 MB/s | 26 kB 00:00 2026-03-09T17:44:28.656 INFO:teuthology.orchestra.run.vm01.stdout:(120/143): python3-kubernetes-26.1.0-3.el9.noar 63 MB/s | 1.0 MB 00:00 2026-03-09T17:44:28.659 INFO:teuthology.orchestra.run.vm01.stdout:(121/143): python3-logutils-0.3.5-21.el9.noarch 17 MB/s | 46 kB 00:00 2026-03-09T17:44:28.663 INFO:teuthology.orchestra.run.vm01.stdout:(122/143): python3-more-itertools-8.12.0-2.el9. 23 MB/s | 79 kB 00:00 2026-03-09T17:44:28.665 INFO:teuthology.orchestra.run.vm01.stdout:(123/143): python3-natsort-7.1.1-5.el9.noarch.r 22 MB/s | 58 kB 00:00 2026-03-09T17:44:28.671 INFO:teuthology.orchestra.run.vm01.stdout:(124/143): python3-pecan-1.4.2-3.el9.noarch.rpm 54 MB/s | 272 kB 00:00 2026-03-09T17:44:28.688 INFO:teuthology.orchestra.run.vm01.stdout:(125/143): python3-numpy-1.23.5-2.el9.x86_64.rp 4.3 MB/s | 6.1 MB 00:01 2026-03-09T17:44:28.693 INFO:teuthology.orchestra.run.vm01.stdout:(126/143): python3-portend-3.1.0-2.el9.noarch.r 757 kB/s | 16 kB 00:00 2026-03-09T17:44:28.695 INFO:teuthology.orchestra.run.vm01.stdout:(127/143): python3-pyOpenSSL-21.0.0-1.el9.noarc 13 MB/s | 90 kB 00:00 2026-03-09T17:44:28.696 INFO:teuthology.orchestra.run.vm01.stdout:(128/143): python3-repoze-lru-0.7-16.el9.noarch 9.1 MB/s | 31 kB 00:00 2026-03-09T17:44:28.701 INFO:teuthology.orchestra.run.vm01.stdout:(129/143): python3-rsa-4.9-2.el9.noarch.rpm 14 MB/s | 59 kB 00:00 2026-03-09T17:44:28.703 INFO:teuthology.orchestra.run.vm01.stdout:(130/143): python3-routes-2.5.1-5.el9.noarch.rp 24 MB/s | 188 kB 00:00 2026-03-09T17:44:28.704 INFO:teuthology.orchestra.run.vm01.stdout:(131/143): python3-tempora-5.0.0-2.el9.noarch.r 10 MB/s | 36 kB 00:00 2026-03-09T17:44:28.707 INFO:teuthology.orchestra.run.vm01.stdout:(132/143): python3-typing-extensions-4.15.0-1.e 23 MB/s | 86 kB 00:00 2026-03-09T17:44:28.711 INFO:teuthology.orchestra.run.vm01.stdout:(133/143): python3-websocket-client-1.2.3-2.el9 24 MB/s | 90 kB 00:00 2026-03-09T17:44:28.713 INFO:teuthology.orchestra.run.vm01.stdout:(134/143): python3-webob-1.8.8-2.el9.noarch.rpm 25 MB/s | 230 kB 00:00 2026-03-09T17:44:28.716 INFO:teuthology.orchestra.run.vm01.stdout:(135/143): python3-xmltodict-0.12.0-15.el9.noar 9.4 MB/s | 22 kB 00:00 2026-03-09T17:44:28.718 INFO:teuthology.orchestra.run.vm01.stdout:(136/143): python3-zc-lockfile-2.0-10.el9.noarc 8.5 MB/s | 20 kB 00:00 2026-03-09T17:44:28.722 INFO:teuthology.orchestra.run.vm01.stdout:(137/143): python3-werkzeug-2.0.3-3.el9.1.noarc 39 MB/s | 427 kB 00:00 2026-03-09T17:44:28.728 INFO:teuthology.orchestra.run.vm01.stdout:(138/143): re2-20211101-20.el9.x86_64.rpm 20 MB/s | 191 kB 00:00 2026-03-09T17:44:28.748 INFO:teuthology.orchestra.run.vm01.stdout:(139/143): thrift-0.15.0-4.el9.x86_64.rpm 61 MB/s | 1.6 MB 00:00 2026-03-09T17:44:28.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:28 vm01 ceph-mon[52793]: pgmap v138: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:29.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:28 vm04 ceph-mon[53158]: pgmap v138: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:29.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:28 vm05 ceph-mon[53831]: pgmap v138: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:29.696 INFO:teuthology.orchestra.run.vm01.stdout:(140/143): librbd1-19.2.3-678.ge911bdeb.el9.x86 3.3 MB/s | 3.2 MB 00:00 2026-03-09T17:44:29.725 INFO:teuthology.orchestra.run.vm01.stdout:(141/143): librados2-19.2.3-678.ge911bdeb.el9.x 3.4 MB/s | 3.4 MB 00:00 2026-03-09T17:44:29.727 INFO:teuthology.orchestra.run.vm01.stdout:(142/143): kpartx-0.8.7-45.el9.x86_64.rpm 1.5 MB/s | 49 kB 00:00 2026-03-09T17:44:29.952 INFO:teuthology.orchestra.run.vm01.stdout:(143/143): python3-scipy-1.9.3-2.el9.x86_64.rpm 8.0 MB/s | 19 MB 00:02 2026-03-09T17:44:29.955 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-09T17:44:29.955 INFO:teuthology.orchestra.run.vm01.stdout:Total 15 MB/s | 212 MB 00:13 2026-03-09T17:44:30.003 INFO:teuthology.orchestra.run.vm04.stdout:(1/143): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 14 kB/s | 6.5 kB 00:00 2026-03-09T17:44:30.560 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T17:44:30.600 INFO:teuthology.orchestra.run.vm05.stdout:lab-extras 65 kB/s | 50 kB 00:00 2026-03-09T17:44:30.613 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T17:44:30.613 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T17:44:30.785 INFO:teuthology.orchestra.run.vm04.stdout:(2/143): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.5 MB/s | 1.2 MB 00:00 2026-03-09T17:44:30.901 INFO:teuthology.orchestra.run.vm04.stdout:(3/143): ceph-immutable-object-cache-19.2.3-678 1.2 MB/s | 145 kB 00:00 2026-03-09T17:44:30.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:30 vm01 ceph-mon[52793]: pgmap v139: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:31.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:30 vm04 ceph-mon[53158]: pgmap v139: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:31.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:30 vm05 ceph-mon[53831]: pgmap v139: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:31.221 INFO:teuthology.orchestra.run.vm04.stdout:(4/143): ceph-base-19.2.3-678.ge911bdeb.el9.x86 3.3 MB/s | 5.5 MB 00:01 2026-03-09T17:44:31.245 INFO:teuthology.orchestra.run.vm04.stdout:(5/143): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 7.1 MB/s | 2.4 MB 00:00 2026-03-09T17:44:31.352 INFO:teuthology.orchestra.run.vm04.stdout:(6/143): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 8.2 MB/s | 1.1 MB 00:00 2026-03-09T17:44:31.467 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T17:44:31.468 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T17:44:31.757 INFO:teuthology.orchestra.run.vm04.stdout:(7/143): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 9.3 MB/s | 4.7 MB 00:00 2026-03-09T17:44:31.971 INFO:teuthology.orchestra.run.vm05.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T17:44:31.971 INFO:teuthology.orchestra.run.vm05.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T17:44:31.976 INFO:teuthology.orchestra.run.vm05.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T17:44:31.976 INFO:teuthology.orchestra.run.vm05.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T17:44:32.006 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout:======================================================================================= 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout:======================================================================================= 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout:Installing: 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: device-mapper-multipath x86_64 0.8.7-45.el9 baseos 156 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: iscsi-initiator-utils x86_64 6.2.1.11-0.git4b3e853.el9 baseos 392 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout:Upgrading: 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: kpartx x86_64 0.8.7-45.el9 baseos 49 k 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T17:44:32.011 INFO:teuthology.orchestra.run.vm05.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout:Installing dependencies: 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: device-mapper-multipath-libs x86_64 0.8.7-45.el9 baseos 289 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: iscsi-initiator-utils-iscsiuio x86_64 6.2.1.11-0.git4b3e853.el9 baseos 81 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: isns-utils-libs x86_64 0.101-4.el9 baseos 101 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T17:44:32.012 INFO:teuthology.orchestra.run.vm05.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout:Installing weak dependencies: 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout:======================================================================================= 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout:Install 140 Packages 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout:Upgrade 3 Packages 2026-03-09T17:44:32.013 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:44:32.014 INFO:teuthology.orchestra.run.vm05.stdout:Total download size: 212 M 2026-03-09T17:44:32.014 INFO:teuthology.orchestra.run.vm05.stdout:Downloading Packages: 2026-03-09T17:44:32.394 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T17:44:32.410 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/146 2026-03-09T17:44:32.423 INFO:teuthology.orchestra.run.vm01.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/146 2026-03-09T17:44:32.497 INFO:teuthology.orchestra.run.vm04.stdout:(8/143): ceph-common-19.2.3-678.ge911bdeb.el9.x 7.4 MB/s | 22 MB 00:02 2026-03-09T17:44:32.600 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/146 2026-03-09T17:44:32.602 INFO:teuthology.orchestra.run.vm01.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-09T17:44:32.611 INFO:teuthology.orchestra.run.vm04.stdout:(9/143): ceph-selinux-19.2.3-678.ge911bdeb.el9. 221 kB/s | 25 kB 00:00 2026-03-09T17:44:32.666 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-09T17:44:32.668 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/146 2026-03-09T17:44:32.699 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/146 2026-03-09T17:44:32.707 INFO:teuthology.orchestra.run.vm04.stdout:(10/143): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 13 MB/s | 17 MB 00:01 2026-03-09T17:44:32.715 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/146 2026-03-09T17:44:32.719 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/146 2026-03-09T17:44:32.721 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/146 2026-03-09T17:44:32.729 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/146 2026-03-09T17:44:32.740 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/146 2026-03-09T17:44:32.742 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-09T17:44:32.778 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-09T17:44:32.780 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/146 2026-03-09T17:44:32.795 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/146 2026-03-09T17:44:32.831 INFO:teuthology.orchestra.run.vm04.stdout:(11/143): libcephfs-devel-19.2.3-678.ge911bdeb. 271 kB/s | 34 kB 00:00 2026-03-09T17:44:32.832 INFO:teuthology.orchestra.run.vm01.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/146 2026-03-09T17:44:32.871 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/146 2026-03-09T17:44:32.877 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/146 2026-03-09T17:44:32.902 INFO:teuthology.orchestra.run.vm01.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/146 2026-03-09T17:44:32.912 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/146 2026-03-09T17:44:32.923 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 18/146 2026-03-09T17:44:32.930 INFO:teuthology.orchestra.run.vm01.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 19/146 2026-03-09T17:44:32.934 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lua-5.4.4-4.el9.x86_64 20/146 2026-03-09T17:44:32.940 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 21/146 2026-03-09T17:44:32.967 INFO:teuthology.orchestra.run.vm04.stdout:(12/143): ceph-radosgw-19.2.3-678.ge911bdeb.el9 8.9 MB/s | 11 MB 00:01 2026-03-09T17:44:32.971 INFO:teuthology.orchestra.run.vm01.stdout: Installing : unzip-6.0-59.el9.x86_64 22/146 2026-03-09T17:44:32.975 INFO:teuthology.orchestra.run.vm04.stdout:(13/143): libcephfs2-19.2.3-678.ge911bdeb.el9.x 6.8 MB/s | 1.0 MB 00:00 2026-03-09T17:44:32.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:32 vm01 ceph-mon[52793]: pgmap v140: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:32.990 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 23/146 2026-03-09T17:44:32.993 INFO:teuthology.orchestra.run.vm05.stdout:(1/143): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 13 kB/s | 6.5 kB 00:00 2026-03-09T17:44:32.995 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 24/146 2026-03-09T17:44:33.003 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 25/146 2026-03-09T17:44:33.006 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 26/146 2026-03-09T17:44:33.039 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 27/146 2026-03-09T17:44:33.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:32 vm04 ceph-mon[53158]: pgmap v140: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:33.047 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 28/146 2026-03-09T17:44:33.057 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 29/146 2026-03-09T17:44:33.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:32 vm05 ceph-mon[53831]: pgmap v140: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:33.072 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 30/146 2026-03-09T17:44:33.082 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 31/146 2026-03-09T17:44:33.090 INFO:teuthology.orchestra.run.vm04.stdout:(14/143): libcephsqlite-19.2.3-678.ge911bdeb.el 1.3 MB/s | 163 kB 00:00 2026-03-09T17:44:33.100 INFO:teuthology.orchestra.run.vm04.stdout:(15/143): librados-devel-19.2.3-678.ge911bdeb.e 1.0 MB/s | 127 kB 00:00 2026-03-09T17:44:33.114 INFO:teuthology.orchestra.run.vm01.stdout: Installing : zip-3.0-35.el9.x86_64 32/146 2026-03-09T17:44:33.120 INFO:teuthology.orchestra.run.vm01.stdout: Installing : luarocks-3.9.2-5.el9.noarch 33/146 2026-03-09T17:44:33.129 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 34/146 2026-03-09T17:44:33.162 INFO:teuthology.orchestra.run.vm01.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 35/146 2026-03-09T17:44:33.212 INFO:teuthology.orchestra.run.vm04.stdout:(16/143): libradosstriper1-19.2.3-678.ge911bdeb 4.0 MB/s | 503 kB 00:00 2026-03-09T17:44:33.232 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 36/146 2026-03-09T17:44:33.249 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 37/146 2026-03-09T17:44:33.256 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rsa-4.9-2.el9.noarch 38/146 2026-03-09T17:44:33.265 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 39/146 2026-03-09T17:44:33.270 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 40/146 2026-03-09T17:44:33.278 INFO:teuthology.orchestra.run.vm01.stdout: Upgrading : kpartx-0.8.7-45.el9.x86_64 41/146 2026-03-09T17:44:33.282 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/146 2026-03-09T17:44:33.300 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/146 2026-03-09T17:44:33.325 INFO:teuthology.orchestra.run.vm04.stdout:(17/143): python3-ceph-argparse-19.2.3-678.ge91 400 kB/s | 45 kB 00:00 2026-03-09T17:44:33.327 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/146 2026-03-09T17:44:33.334 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/146 2026-03-09T17:44:33.341 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/146 2026-03-09T17:44:33.357 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/146 2026-03-09T17:44:33.370 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/146 2026-03-09T17:44:33.382 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/146 2026-03-09T17:44:33.439 INFO:teuthology.orchestra.run.vm04.stdout:(18/143): python3-ceph-common-19.2.3-678.ge911b 1.2 MB/s | 142 kB 00:00 2026-03-09T17:44:33.456 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/146 2026-03-09T17:44:33.464 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/146 2026-03-09T17:44:33.474 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/146 2026-03-09T17:44:33.485 INFO:teuthology.orchestra.run.vm04.stdout:(19/143): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 14 MB/s | 5.4 MB 00:00 2026-03-09T17:44:33.534 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/146 2026-03-09T17:44:33.553 INFO:teuthology.orchestra.run.vm04.stdout:(20/143): python3-cephfs-19.2.3-678.ge911bdeb.e 1.4 MB/s | 165 kB 00:00 2026-03-09T17:44:33.608 INFO:teuthology.orchestra.run.vm04.stdout:(21/143): python3-rados-19.2.3-678.ge911bdeb.el 2.6 MB/s | 323 kB 00:00 2026-03-09T17:44:33.669 INFO:teuthology.orchestra.run.vm04.stdout:(22/143): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.6 MB/s | 303 kB 00:00 2026-03-09T17:44:33.729 INFO:teuthology.orchestra.run.vm04.stdout:(23/143): python3-rgw-19.2.3-678.ge911bdeb.el9. 824 kB/s | 100 kB 00:00 2026-03-09T17:44:33.783 INFO:teuthology.orchestra.run.vm04.stdout:(24/143): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 752 kB/s | 85 kB 00:00 2026-03-09T17:44:33.917 INFO:teuthology.orchestra.run.vm05.stdout:(2/143): ceph-base-19.2.3-678.ge911bdeb.el9.x86 3.9 MB/s | 5.5 MB 00:01 2026-03-09T17:44:33.939 INFO:teuthology.orchestra.run.vm04.stdout:(25/143): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.1 MB/s | 171 kB 00:00 2026-03-09T17:44:33.942 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/146 2026-03-09T17:44:33.960 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/146 2026-03-09T17:44:33.966 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/146 2026-03-09T17:44:33.974 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/146 2026-03-09T17:44:33.979 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/146 2026-03-09T17:44:33.987 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/146 2026-03-09T17:44:33.992 INFO:teuthology.orchestra.run.vm01.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/146 2026-03-09T17:44:33.995 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/146 2026-03-09T17:44:34.007 INFO:teuthology.orchestra.run.vm04.stdout:(26/143): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 11 MB/s | 3.1 MB 00:00 2026-03-09T17:44:34.026 INFO:teuthology.orchestra.run.vm01.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/146 2026-03-09T17:44:34.028 INFO:teuthology.orchestra.run.vm05.stdout:(3/143): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.1 MB/s | 1.2 MB 00:01 2026-03-09T17:44:34.031 INFO:teuthology.orchestra.run.vm05.stdout:(4/143): ceph-immutable-object-cache-19.2.3-678 1.2 MB/s | 145 kB 00:00 2026-03-09T17:44:34.052 INFO:teuthology.orchestra.run.vm04.stdout:(27/143): ceph-grafana-dashboards-19.2.3-678.ge 277 kB/s | 31 kB 00:00 2026-03-09T17:44:34.081 INFO:teuthology.orchestra.run.vm01.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/146 2026-03-09T17:44:34.094 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/146 2026-03-09T17:44:34.103 INFO:teuthology.orchestra.run.vm01.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/146 2026-03-09T17:44:34.109 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/146 2026-03-09T17:44:34.116 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/146 2026-03-09T17:44:34.122 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/146 2026-03-09T17:44:34.128 INFO:teuthology.orchestra.run.vm04.stdout:(28/143): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.2 MB/s | 150 kB 00:00 2026-03-09T17:44:34.132 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/146 2026-03-09T17:44:34.138 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/146 2026-03-09T17:44:34.156 INFO:teuthology.orchestra.run.vm05.stdout:(5/143): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 8.6 MB/s | 1.1 MB 00:00 2026-03-09T17:44:34.173 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/146 2026-03-09T17:44:34.186 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/146 2026-03-09T17:44:34.235 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/146 2026-03-09T17:44:34.528 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/146 2026-03-09T17:44:34.563 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/146 2026-03-09T17:44:34.571 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/146 2026-03-09T17:44:34.638 INFO:teuthology.orchestra.run.vm01.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/146 2026-03-09T17:44:34.642 INFO:teuthology.orchestra.run.vm01.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/146 2026-03-09T17:44:34.670 INFO:teuthology.orchestra.run.vm05.stdout:(6/143): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 9.2 MB/s | 4.7 MB 00:00 2026-03-09T17:44:34.676 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/146 2026-03-09T17:44:34.706 INFO:teuthology.orchestra.run.vm05.stdout:(7/143): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 3.6 MB/s | 2.4 MB 00:00 2026-03-09T17:44:34.751 INFO:teuthology.orchestra.run.vm04.stdout:(29/143): ceph-mgr-diskprediction-local-19.2.3- 12 MB/s | 7.4 MB 00:00 2026-03-09T17:44:34.771 INFO:teuthology.orchestra.run.vm04.stdout:(30/143): ceph-mgr-dashboard-19.2.3-678.ge911bd 5.3 MB/s | 3.8 MB 00:00 2026-03-09T17:44:34.873 INFO:teuthology.orchestra.run.vm04.stdout:(31/143): ceph-mgr-modules-core-19.2.3-678.ge91 2.0 MB/s | 253 kB 00:00 2026-03-09T17:44:34.909 INFO:teuthology.orchestra.run.vm05.stdout:(8/143): ceph-common-19.2.3-678.ge911bdeb.el9.x 9.1 MB/s | 22 MB 00:02 2026-03-09T17:44:34.909 INFO:teuthology.orchestra.run.vm04.stdout:(32/143): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 357 kB/s | 49 kB 00:00 2026-03-09T17:44:34.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:34 vm01 ceph-mon[52793]: pgmap v141: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:34.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:44:34 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:44:34.993 INFO:teuthology.orchestra.run.vm04.stdout:(33/143): ceph-prometheus-alerts-19.2.3-678.ge9 139 kB/s | 17 kB 00:00 2026-03-09T17:44:35.030 INFO:teuthology.orchestra.run.vm04.stdout:(34/143): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 299 kB 00:00 2026-03-09T17:44:35.035 INFO:teuthology.orchestra.run.vm05.stdout:(9/143): ceph-selinux-19.2.3-678.ge911bdeb.el9. 199 kB/s | 25 kB 00:00 2026-03-09T17:44:35.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:34 vm04 ceph-mon[53158]: pgmap v141: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:35.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:34 vm05 ceph-mon[53831]: pgmap v141: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:35.085 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/146 2026-03-09T17:44:35.123 INFO:teuthology.orchestra.run.vm04.stdout:(35/143): cephadm-19.2.3-678.ge911bdeb.el9.noar 5.8 MB/s | 769 kB 00:00 2026-03-09T17:44:35.180 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/146 2026-03-09T17:44:35.252 INFO:teuthology.orchestra.run.vm04.stdout:(36/143): cryptsetup-2.8.1-3.el9.x86_64.rpm 1.5 MB/s | 351 kB 00:00 2026-03-09T17:44:35.288 INFO:teuthology.orchestra.run.vm04.stdout:(37/143): device-mapper-multipath-0.8.7-45.el9. 948 kB/s | 156 kB 00:00 2026-03-09T17:44:35.317 INFO:teuthology.orchestra.run.vm04.stdout:(38/143): device-mapper-multipath-libs-0.8.7-45 4.4 MB/s | 289 kB 00:00 2026-03-09T17:44:35.376 INFO:teuthology.orchestra.run.vm04.stdout:(39/143): iscsi-initiator-utils-iscsiuio-6.2.1. 1.4 MB/s | 81 kB 00:00 2026-03-09T17:44:35.444 INFO:teuthology.orchestra.run.vm04.stdout:(40/143): iscsi-initiator-utils-6.2.1.11-0.git4 2.5 MB/s | 392 kB 00:00 2026-03-09T17:44:35.455 INFO:teuthology.orchestra.run.vm04.stdout:(41/143): isns-utils-libs-0.101-4.el9.x86_64.rp 1.2 MB/s | 101 kB 00:00 2026-03-09T17:44:35.483 INFO:teuthology.orchestra.run.vm04.stdout:(42/143): ledmon-libs-1.1.0-3.el9.x86_64.rpm 1.0 MB/s | 40 kB 00:00 2026-03-09T17:44:35.515 INFO:teuthology.orchestra.run.vm04.stdout:(43/143): libconfig-1.7.2-9.el9.x86_64.rpm 1.2 MB/s | 72 kB 00:00 2026-03-09T17:44:35.736 INFO:teuthology.orchestra.run.vm04.stdout:(44/143): libquadmath-11.5.0-14.el9.x86_64.rpm 837 kB/s | 184 kB 00:00 2026-03-09T17:44:35.770 INFO:teuthology.orchestra.run.vm04.stdout:(45/143): mailcap-2.1.49-5.el9.noarch.rpm 1.0 MB/s | 33 kB 00:00 2026-03-09T17:44:35.826 INFO:teuthology.orchestra.run.vm04.stdout:(46/143): libgfortran-11.5.0-14.el9.x86_64.rpm 2.3 MB/s | 794 kB 00:00 2026-03-09T17:44:35.834 INFO:teuthology.orchestra.run.vm04.stdout:(47/143): pciutils-3.7.0-7.el9.x86_64.rpm 1.4 MB/s | 93 kB 00:00 2026-03-09T17:44:35.948 INFO:teuthology.orchestra.run.vm04.stdout:(48/143): python3-cryptography-36.0.1-5.el9.x86 11 MB/s | 1.2 MB 00:00 2026-03-09T17:44:35.957 INFO:teuthology.orchestra.run.vm04.stdout:(49/143): python3-cffi-1.14.5-5.el9.x86_64.rpm 1.9 MB/s | 253 kB 00:00 2026-03-09T17:44:35.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:35 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:36.001 INFO:teuthology.orchestra.run.vm04.stdout:(50/143): python3-ply-3.11-14.el9.noarch.rpm 2.0 MB/s | 106 kB 00:00 2026-03-09T17:44:36.008 INFO:teuthology.orchestra.run.vm04.stdout:(51/143): python3-pycparser-2.20-6.el9.noarch.r 2.6 MB/s | 135 kB 00:00 2026-03-09T17:44:36.035 INFO:teuthology.orchestra.run.vm04.stdout:(52/143): python3-pyparsing-2.4.7-9.el9.noarch. 4.3 MB/s | 150 kB 00:00 2026-03-09T17:44:36.037 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/146 2026-03-09T17:44:36.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:35 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:36.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:35 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:36.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:44:35 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:44:36.073 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/146 2026-03-09T17:44:36.080 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/146 2026-03-09T17:44:36.085 INFO:teuthology.orchestra.run.vm01.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/146 2026-03-09T17:44:36.095 INFO:teuthology.orchestra.run.vm04.stdout:(53/143): python3-requests-2.25.1-10.el9.noarch 1.4 MB/s | 126 kB 00:00 2026-03-09T17:44:36.117 INFO:teuthology.orchestra.run.vm04.stdout:(54/143): python3-urllib3-1.26.5-7.el9.noarch.r 2.6 MB/s | 218 kB 00:00 2026-03-09T17:44:36.155 INFO:teuthology.orchestra.run.vm05.stdout:(10/143): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 11 MB/s | 17 MB 00:01 2026-03-09T17:44:36.176 INFO:teuthology.orchestra.run.vm04.stdout:(55/143): zip-3.0-35.el9.x86_64.rpm 4.4 MB/s | 266 kB 00:00 2026-03-09T17:44:36.193 INFO:teuthology.orchestra.run.vm04.stdout:(56/143): unzip-6.0-59.el9.x86_64.rpm 1.8 MB/s | 182 kB 00:00 2026-03-09T17:44:36.247 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/146 2026-03-09T17:44:36.250 INFO:teuthology.orchestra.run.vm01.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/146 2026-03-09T17:44:36.264 INFO:teuthology.orchestra.run.vm05.stdout:(11/143): libcephfs-devel-19.2.3-678.ge911bdeb. 309 kB/s | 34 kB 00:00 2026-03-09T17:44:36.284 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/146 2026-03-09T17:44:36.288 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/146 2026-03-09T17:44:36.291 INFO:teuthology.orchestra.run.vm04.stdout:(57/143): flexiblas-3.0.4-9.el9.x86_64.rpm 303 kB/s | 30 kB 00:00 2026-03-09T17:44:36.300 INFO:teuthology.orchestra.run.vm01.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/146 2026-03-09T17:44:36.304 INFO:teuthology.orchestra.run.vm04.stdout:(58/143): boost-program-options-1.75.0-13.el9.x 816 kB/s | 104 kB 00:00 2026-03-09T17:44:36.394 INFO:teuthology.orchestra.run.vm05.stdout:(12/143): libcephfs2-19.2.3-678.ge911bdeb.el9.x 7.6 MB/s | 1.0 MB 00:00 2026-03-09T17:44:36.406 INFO:teuthology.orchestra.run.vm04.stdout:(59/143): flexiblas-openblas-openmp-3.0.4-9.el9 145 kB/s | 15 kB 00:00 2026-03-09T17:44:36.464 INFO:teuthology.orchestra.run.vm04.stdout:(60/143): libnbd-1.20.3-4.el9.x86_64.rpm 2.8 MB/s | 164 kB 00:00 2026-03-09T17:44:36.481 INFO:teuthology.orchestra.run.vm04.stdout:(61/143): flexiblas-netlib-3.0.4-9.el9.x86_64.r 16 MB/s | 3.0 MB 00:00 2026-03-09T17:44:36.504 INFO:teuthology.orchestra.run.vm05.stdout:(13/143): libcephsqlite-19.2.3-678.ge911bdeb.el 1.4 MB/s | 163 kB 00:00 2026-03-09T17:44:36.516 INFO:teuthology.orchestra.run.vm04.stdout:(62/143): libpmemobj-1.12.1-1.el9.x86_64.rpm 3.0 MB/s | 160 kB 00:00 2026-03-09T17:44:36.526 INFO:teuthology.orchestra.run.vm04.stdout:(63/143): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.0 MB/s | 45 kB 00:00 2026-03-09T17:44:36.565 INFO:teuthology.orchestra.run.vm01.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/146 2026-03-09T17:44:36.569 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/146 2026-03-09T17:44:36.574 INFO:teuthology.orchestra.run.vm04.stdout:(64/143): libstoragemgmt-1.10.1-1.el9.x86_64.rp 5.1 MB/s | 246 kB 00:00 2026-03-09T17:44:36.582 INFO:teuthology.orchestra.run.vm04.stdout:(65/143): librdkafka-1.6.1-102.el9.x86_64.rpm 10 MB/s | 662 kB 00:00 2026-03-09T17:44:36.592 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/146 2026-03-09T17:44:36.595 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/146 2026-03-09T17:44:36.615 INFO:teuthology.orchestra.run.vm05.stdout:(14/143): librados-devel-19.2.3-678.ge911bdeb.e 1.1 MB/s | 127 kB 00:00 2026-03-09T17:44:36.625 INFO:teuthology.orchestra.run.vm04.stdout:(66/143): libxslt-1.1.34-12.el9.x86_64.rpm 4.5 MB/s | 233 kB 00:00 2026-03-09T17:44:36.632 INFO:teuthology.orchestra.run.vm04.stdout:(67/143): lttng-ust-2.12.0-6.el9.x86_64.rpm 5.7 MB/s | 292 kB 00:00 2026-03-09T17:44:36.672 INFO:teuthology.orchestra.run.vm04.stdout:(68/143): lua-5.4.4-4.el9.x86_64.rpm 3.9 MB/s | 188 kB 00:00 2026-03-09T17:44:36.674 INFO:teuthology.orchestra.run.vm04.stdout:(69/143): openblas-0.3.29-1.el9.x86_64.rpm 1.0 MB/s | 42 kB 00:00 2026-03-09T17:44:36.793 INFO:teuthology.orchestra.run.vm04.stdout:(70/143): protobuf-3.14.0-17.el9.x86_64.rpm 8.5 MB/s | 1.0 MB 00:00 2026-03-09T17:44:36.856 INFO:teuthology.orchestra.run.vm04.stdout:(71/143): openblas-openmp-0.3.29-1.el9.x86_64.r 29 MB/s | 5.3 MB 00:00 2026-03-09T17:44:36.932 INFO:teuthology.orchestra.run.vm04.stdout:(72/143): python3-devel-3.9.25-3.el9.x86_64.rpm 3.2 MB/s | 244 kB 00:00 2026-03-09T17:44:36.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:36 vm01 ceph-mon[52793]: pgmap v142: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:36.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:36 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:36.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:36 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:44:37.006 INFO:teuthology.orchestra.run.vm04.stdout:(73/143): python3-babel-2.9.1-2.el9.noarch.rpm 28 MB/s | 6.0 MB 00:00 2026-03-09T17:44:37.007 INFO:teuthology.orchestra.run.vm04.stdout:(74/143): python3-jinja2-2.11.3-8.el9.noarch.rp 3.2 MB/s | 249 kB 00:00 2026-03-09T17:44:37.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:36 vm04 ceph-mon[53158]: pgmap v142: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:37.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:36 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:37.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:36 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:44:37.043 INFO:teuthology.orchestra.run.vm05.stdout:(15/143): ceph-radosgw-19.2.3-678.ge911bdeb.el9 4.6 MB/s | 11 MB 00:02 2026-03-09T17:44:37.057 INFO:teuthology.orchestra.run.vm04.stdout:(75/143): python3-libstoragemgmt-1.10.1-1.el9.x 3.5 MB/s | 177 kB 00:00 2026-03-09T17:44:37.058 INFO:teuthology.orchestra.run.vm04.stdout:(76/143): python3-jmespath-1.0.1-1.el9.noarch.r 951 kB/s | 48 kB 00:00 2026-03-09T17:44:37.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:36 vm05 ceph-mon[53831]: pgmap v142: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:37.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:36 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:37.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:36 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:44:37.109 INFO:teuthology.orchestra.run.vm04.stdout:(77/143): python3-mako-1.1.4-6.el9.noarch.rpm 3.3 MB/s | 172 kB 00:00 2026-03-09T17:44:37.117 INFO:teuthology.orchestra.run.vm04.stdout:(78/143): python3-markupsafe-1.1.1-12.el9.x86_6 584 kB/s | 35 kB 00:00 2026-03-09T17:44:37.215 INFO:teuthology.orchestra.run.vm04.stdout:(79/143): python3-numpy-f2py-1.23.5-2.el9.x86_6 4.4 MB/s | 442 kB 00:00 2026-03-09T17:44:37.271 INFO:teuthology.orchestra.run.vm04.stdout:(80/143): python3-numpy-1.23.5-2.el9.x86_64.rpm 38 MB/s | 6.1 MB 00:00 2026-03-09T17:44:37.273 INFO:teuthology.orchestra.run.vm04.stdout:(81/143): python3-packaging-20.9-5.el9.noarch.r 1.3 MB/s | 77 kB 00:00 2026-03-09T17:44:37.324 INFO:teuthology.orchestra.run.vm04.stdout:(82/143): python3-pyasn1-0.4.8-7.el9.noarch.rpm 3.0 MB/s | 157 kB 00:00 2026-03-09T17:44:37.329 INFO:teuthology.orchestra.run.vm04.stdout:(83/143): python3-protobuf-3.14.0-17.el9.noarch 4.6 MB/s | 267 kB 00:00 2026-03-09T17:44:37.374 INFO:teuthology.orchestra.run.vm04.stdout:(84/143): python3-pyasn1-modules-0.4.8-7.el9.no 5.4 MB/s | 277 kB 00:00 2026-03-09T17:44:37.379 INFO:teuthology.orchestra.run.vm04.stdout:(85/143): python3-requests-oauthlib-1.3.0-12.el 1.1 MB/s | 54 kB 00:00 2026-03-09T17:44:37.441 INFO:teuthology.orchestra.run.vm04.stdout:(86/143): python3-toml-0.10.2-6.el9.noarch.rpm 674 kB/s | 42 kB 00:00 2026-03-09T17:44:37.562 INFO:teuthology.orchestra.run.vm04.stdout:(87/143): qatlib-25.08.0-2.el9.x86_64.rpm 1.9 MB/s | 240 kB 00:00 2026-03-09T17:44:37.646 INFO:teuthology.orchestra.run.vm04.stdout:(88/143): qatlib-service-25.08.0-2.el9.x86_64.r 447 kB/s | 37 kB 00:00 2026-03-09T17:44:37.734 INFO:teuthology.orchestra.run.vm04.stdout:(89/143): python3-scipy-1.9.3-2.el9.x86_64.rpm 54 MB/s | 19 MB 00:00 2026-03-09T17:44:37.736 INFO:teuthology.orchestra.run.vm04.stdout:(90/143): qatzip-libs-1.3.1-1.el9.x86_64.rpm 745 kB/s | 66 kB 00:00 2026-03-09T17:44:37.764 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-09T17:44:37.769 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-09T17:44:37.783 INFO:teuthology.orchestra.run.vm04.stdout:(91/143): socat-1.7.4.1-8.el9.x86_64.rpm 6.1 MB/s | 303 kB 00:00 2026-03-09T17:44:37.784 INFO:teuthology.orchestra.run.vm04.stdout:(92/143): xmlstarlet-1.6.1-20.el9.x86_64.rpm 1.3 MB/s | 64 kB 00:00 2026-03-09T17:44:37.794 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-09T17:44:37.807 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 94/146 2026-03-09T17:44:37.816 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-packaging-20.9-5.el9.noarch 95/146 2026-03-09T17:44:37.836 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ply-3.11-14.el9.noarch 96/146 2026-03-09T17:44:37.919 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 97/146 2026-03-09T17:44:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:44:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:44:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:44:38.019 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 98/146 2026-03-09T17:44:38.028 INFO:teuthology.orchestra.run.vm04.stdout:(93/143): lua-devel-5.4.4-4.el9.x86_64.rpm 91 kB/s | 22 kB 00:00 2026-03-09T17:44:38.035 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 99/146 2026-03-09T17:44:38.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:37 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:44:38.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:37 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:44:38.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:37 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:44:38.043 INFO:teuthology.orchestra.run.vm04.stdout:(94/143): abseil-cpp-20211102.0-4.el9.x86_64.rp 37 MB/s | 551 kB 00:00 2026-03-09T17:44:38.050 INFO:teuthology.orchestra.run.vm04.stdout:(95/143): gperftools-libs-2.9.1-3.el9.x86_64.rp 42 MB/s | 308 kB 00:00 2026-03-09T17:44:38.053 INFO:teuthology.orchestra.run.vm04.stdout:(96/143): grpc-data-1.46.7-10.el9.noarch.rpm 8.0 MB/s | 19 kB 00:00 2026-03-09T17:44:38.066 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 100/146 2026-03-09T17:44:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:37 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:44:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:37 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:44:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:37 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:44:38.107 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 101/146 2026-03-09T17:44:38.113 INFO:teuthology.orchestra.run.vm04.stdout:(97/143): libarrow-9.0.0-15.el9.x86_64.rpm 74 MB/s | 4.4 MB 00:00 2026-03-09T17:44:38.113 INFO:teuthology.orchestra.run.vm05.stdout:(16/143): libradosstriper1-19.2.3-678.ge911bdeb 336 kB/s | 503 kB 00:01 2026-03-09T17:44:38.116 INFO:teuthology.orchestra.run.vm04.stdout:(98/143): libarrow-doc-9.0.0-15.el9.noarch.rpm 9.2 MB/s | 25 kB 00:00 2026-03-09T17:44:38.118 INFO:teuthology.orchestra.run.vm04.stdout:(99/143): liboath-2.6.12-1.el9.x86_64.rpm 19 MB/s | 49 kB 00:00 2026-03-09T17:44:38.121 INFO:teuthology.orchestra.run.vm04.stdout:(100/143): libunwind-1.6.2-1.el9.x86_64.rpm 24 MB/s | 67 kB 00:00 2026-03-09T17:44:38.125 INFO:teuthology.orchestra.run.vm04.stdout:(101/143): luarocks-3.9.2-5.el9.noarch.rpm 42 MB/s | 151 kB 00:00 2026-03-09T17:44:38.142 INFO:teuthology.orchestra.run.vm04.stdout:(102/143): parquet-libs-9.0.0-15.el9.x86_64.rpm 48 MB/s | 838 kB 00:00 2026-03-09T17:44:38.152 INFO:teuthology.orchestra.run.vm04.stdout:(103/143): python3-asyncssh-2.13.2-5.el9.noarch 58 MB/s | 548 kB 00:00 2026-03-09T17:44:38.155 INFO:teuthology.orchestra.run.vm04.stdout:(104/143): python3-autocommand-2.2.2-8.el9.noar 11 MB/s | 29 kB 00:00 2026-03-09T17:44:38.157 INFO:teuthology.orchestra.run.vm04.stdout:(105/143): python3-backports-tarfile-1.2.0-1.el 23 MB/s | 60 kB 00:00 2026-03-09T17:44:38.160 INFO:teuthology.orchestra.run.vm04.stdout:(106/143): python3-bcrypt-3.2.2-1.el9.x86_64.rp 17 MB/s | 43 kB 00:00 2026-03-09T17:44:38.163 INFO:teuthology.orchestra.run.vm04.stdout:(107/143): python3-cachetools-4.2.4-1.el9.noarc 12 MB/s | 32 kB 00:00 2026-03-09T17:44:38.165 INFO:teuthology.orchestra.run.vm04.stdout:(108/143): python3-certifi-2023.05.07-4.el9.noa 6.7 MB/s | 14 kB 00:00 2026-03-09T17:44:38.168 INFO:teuthology.orchestra.run.vm04.stdout:(109/143): python3-cheroot-10.0.1-4.el9.noarch. 48 MB/s | 173 kB 00:00 2026-03-09T17:44:38.175 INFO:teuthology.orchestra.run.vm04.stdout:(110/143): python3-cherrypy-18.6.1-2.el9.noarch 58 MB/s | 358 kB 00:00 2026-03-09T17:44:38.180 INFO:teuthology.orchestra.run.vm04.stdout:(111/143): python3-google-auth-2.45.0-1.el9.noa 52 MB/s | 254 kB 00:00 2026-03-09T17:44:38.182 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 102/146 2026-03-09T17:44:38.192 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 103/146 2026-03-09T17:44:38.199 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 104/146 2026-03-09T17:44:38.207 INFO:teuthology.orchestra.run.vm01.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 105/146 2026-03-09T17:44:38.209 INFO:teuthology.orchestra.run.vm04.stdout:(112/143): python3-grpcio-1.46.7-10.el9.x86_64. 70 MB/s | 2.0 MB 00:00 2026-03-09T17:44:38.212 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 106/146 2026-03-09T17:44:38.213 INFO:teuthology.orchestra.run.vm04.stdout:(113/143): python3-grpcio-tools-1.46.7-10.el9.x 41 MB/s | 144 kB 00:00 2026-03-09T17:44:38.214 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 107/146 2026-03-09T17:44:38.215 INFO:teuthology.orchestra.run.vm04.stdout:(114/143): python3-jaraco-8.2.1-3.el9.noarch.rp 5.5 MB/s | 11 kB 00:00 2026-03-09T17:44:38.217 INFO:teuthology.orchestra.run.vm04.stdout:(115/143): python3-jaraco-classes-3.2.1-5.el9.n 8.5 MB/s | 18 kB 00:00 2026-03-09T17:44:38.219 INFO:teuthology.orchestra.run.vm04.stdout:(116/143): python3-jaraco-collections-3.0.0-8.e 11 MB/s | 23 kB 00:00 2026-03-09T17:44:38.221 INFO:teuthology.orchestra.run.vm04.stdout:(117/143): python3-jaraco-context-6.0.1-3.el9.n 10 MB/s | 20 kB 00:00 2026-03-09T17:44:38.223 INFO:teuthology.orchestra.run.vm05.stdout:(17/143): python3-ceph-argparse-19.2.3-678.ge91 412 kB/s | 45 kB 00:00 2026-03-09T17:44:38.224 INFO:teuthology.orchestra.run.vm04.stdout:(118/143): python3-jaraco-functools-3.5.0-2.el9 9.0 MB/s | 19 kB 00:00 2026-03-09T17:44:38.226 INFO:teuthology.orchestra.run.vm04.stdout:(119/143): python3-jaraco-text-4.0.0-2.el9.noar 11 MB/s | 26 kB 00:00 2026-03-09T17:44:38.232 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 107/146 2026-03-09T17:44:38.241 INFO:teuthology.orchestra.run.vm04.stdout:(120/143): python3-kubernetes-26.1.0-3.el9.noar 71 MB/s | 1.0 MB 00:00 2026-03-09T17:44:38.244 INFO:teuthology.orchestra.run.vm04.stdout:(121/143): python3-logutils-0.3.5-21.el9.noarch 16 MB/s | 46 kB 00:00 2026-03-09T17:44:38.247 INFO:teuthology.orchestra.run.vm04.stdout:(122/143): python3-more-itertools-8.12.0-2.el9. 27 MB/s | 79 kB 00:00 2026-03-09T17:44:38.250 INFO:teuthology.orchestra.run.vm04.stdout:(123/143): python3-natsort-7.1.1-5.el9.noarch.r 22 MB/s | 58 kB 00:00 2026-03-09T17:44:38.255 INFO:teuthology.orchestra.run.vm04.stdout:(124/143): python3-pecan-1.4.2-3.el9.noarch.rpm 50 MB/s | 272 kB 00:00 2026-03-09T17:44:38.260 INFO:teuthology.orchestra.run.vm04.stdout:(125/143): python3-portend-3.1.0-2.el9.noarch.r 3.5 MB/s | 16 kB 00:00 2026-03-09T17:44:38.264 INFO:teuthology.orchestra.run.vm04.stdout:(126/143): python3-pyOpenSSL-21.0.0-1.el9.noarc 23 MB/s | 90 kB 00:00 2026-03-09T17:44:38.266 INFO:teuthology.orchestra.run.vm04.stdout:(127/143): python3-repoze-lru-0.7-16.el9.noarch 14 MB/s | 31 kB 00:00 2026-03-09T17:44:38.271 INFO:teuthology.orchestra.run.vm04.stdout:(128/143): python3-routes-2.5.1-5.el9.noarch.rp 42 MB/s | 188 kB 00:00 2026-03-09T17:44:38.273 INFO:teuthology.orchestra.run.vm04.stdout:(129/143): python3-rsa-4.9-2.el9.noarch.rpm 21 MB/s | 59 kB 00:00 2026-03-09T17:44:38.276 INFO:teuthology.orchestra.run.vm04.stdout:(130/143): python3-tempora-5.0.0-2.el9.noarch.r 15 MB/s | 36 kB 00:00 2026-03-09T17:44:38.279 INFO:teuthology.orchestra.run.vm04.stdout:(131/143): python3-typing-extensions-4.15.0-1.e 29 MB/s | 86 kB 00:00 2026-03-09T17:44:38.284 INFO:teuthology.orchestra.run.vm04.stdout:(132/143): python3-webob-1.8.8-2.el9.noarch.rpm 43 MB/s | 230 kB 00:00 2026-03-09T17:44:38.288 INFO:teuthology.orchestra.run.vm04.stdout:(133/143): python3-websocket-client-1.2.3-2.el9 27 MB/s | 90 kB 00:00 2026-03-09T17:44:38.296 INFO:teuthology.orchestra.run.vm04.stdout:(134/143): python3-werkzeug-2.0.3-3.el9.1.noarc 56 MB/s | 427 kB 00:00 2026-03-09T17:44:38.298 INFO:teuthology.orchestra.run.vm04.stdout:(135/143): python3-xmltodict-0.12.0-15.el9.noar 8.7 MB/s | 22 kB 00:00 2026-03-09T17:44:38.301 INFO:teuthology.orchestra.run.vm04.stdout:(136/143): python3-zc-lockfile-2.0-10.el9.noarc 8.6 MB/s | 20 kB 00:00 2026-03-09T17:44:38.305 INFO:teuthology.orchestra.run.vm04.stdout:(137/143): re2-20211101-20.el9.x86_64.rpm 46 MB/s | 191 kB 00:00 2026-03-09T17:44:38.327 INFO:teuthology.orchestra.run.vm04.stdout:(138/143): thrift-0.15.0-4.el9.x86_64.rpm 73 MB/s | 1.6 MB 00:00 2026-03-09T17:44:38.333 INFO:teuthology.orchestra.run.vm05.stdout:(18/143): python3-ceph-common-19.2.3-678.ge911b 1.3 MB/s | 142 kB 00:00 2026-03-09T17:44:38.381 INFO:teuthology.orchestra.run.vm04.stdout:(139/143): protobuf-compiler-3.14.0-17.el9.x86_ 1.4 MB/s | 862 kB 00:00 2026-03-09T17:44:38.444 INFO:teuthology.orchestra.run.vm05.stdout:(19/143): python3-cephfs-19.2.3-678.ge911bdeb.e 1.5 MB/s | 165 kB 00:00 2026-03-09T17:44:38.547 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 108/146 2026-03-09T17:44:38.553 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 109/146 2026-03-09T17:44:38.565 INFO:teuthology.orchestra.run.vm05.stdout:(20/143): python3-rados-19.2.3-678.ge911bdeb.el 2.6 MB/s | 323 kB 00:00 2026-03-09T17:44:38.596 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 109/146 2026-03-09T17:44:38.597 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T17:44:38.597 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:38.600 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 110/146 2026-03-09T17:44:38.677 INFO:teuthology.orchestra.run.vm05.stdout:(21/143): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.6 MB/s | 303 kB 00:00 2026-03-09T17:44:38.787 INFO:teuthology.orchestra.run.vm05.stdout:(22/143): python3-rgw-19.2.3-678.ge911bdeb.el9. 908 kB/s | 100 kB 00:00 2026-03-09T17:44:38.898 INFO:teuthology.orchestra.run.vm05.stdout:(23/143): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 773 kB/s | 85 kB 00:00 2026-03-09T17:44:38.911 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:38 vm01 ceph-mon[52793]: pgmap v143: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:38.911 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:38 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1833394149' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-09T17:44:39.109 INFO:teuthology.orchestra.run.vm05.stdout:(24/143): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 2.6 MB/s | 5.4 MB 00:02 2026-03-09T17:44:39.262 INFO:teuthology.orchestra.run.vm05.stdout:(25/143): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.1 MB/s | 171 kB 00:00 2026-03-09T17:44:39.273 INFO:teuthology.orchestra.run.vm05.stdout:(26/143): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 8.3 MB/s | 3.1 MB 00:00 2026-03-09T17:44:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:38 vm04 ceph-mon[53158]: pgmap v143: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:38 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1833394149' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-09T17:44:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:38 vm05 ceph-mon[53831]: pgmap v143: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:38 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1833394149' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-09T17:44:39.347 INFO:teuthology.orchestra.run.vm04.stdout:(140/143): librbd1-19.2.3-678.ge911bdeb.el9.x86 3.3 MB/s | 3.2 MB 00:00 2026-03-09T17:44:39.379 INFO:teuthology.orchestra.run.vm04.stdout:(141/143): kpartx-0.8.7-45.el9.x86_64.rpm 1.5 MB/s | 49 kB 00:00 2026-03-09T17:44:39.385 INFO:teuthology.orchestra.run.vm05.stdout:(27/143): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.3 MB/s | 150 kB 00:00 2026-03-09T17:44:39.386 INFO:teuthology.orchestra.run.vm05.stdout:(28/143): ceph-grafana-dashboards-19.2.3-678.ge 251 kB/s | 31 kB 00:00 2026-03-09T17:44:39.454 INFO:teuthology.orchestra.run.vm04.stdout:(142/143): librados2-19.2.3-678.ge911bdeb.el9.x 3.0 MB/s | 3.4 MB 00:01 2026-03-09T17:44:40.045 INFO:teuthology.orchestra.run.vm04.stdout:(143/143): ceph-test-19.2.3-678.ge911bdeb.el9.x 6.7 MB/s | 50 MB 00:07 2026-03-09T17:44:40.050 INFO:teuthology.orchestra.run.vm04.stdout:-------------------------------------------------------------------------------- 2026-03-09T17:44:40.050 INFO:teuthology.orchestra.run.vm04.stdout:Total 17 MB/s | 212 MB 00:12 2026-03-09T17:44:40.166 INFO:teuthology.orchestra.run.vm05.stdout:(29/143): ceph-mgr-diskprediction-local-19.2.3- 9.5 MB/s | 7.4 MB 00:00 2026-03-09T17:44:40.286 INFO:teuthology.orchestra.run.vm05.stdout:(30/143): ceph-mgr-modules-core-19.2.3-678.ge91 2.1 MB/s | 253 kB 00:00 2026-03-09T17:44:40.396 INFO:teuthology.orchestra.run.vm05.stdout:(31/143): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 450 kB/s | 49 kB 00:00 2026-03-09T17:44:40.505 INFO:teuthology.orchestra.run.vm05.stdout:(32/143): ceph-prometheus-alerts-19.2.3-678.ge9 154 kB/s | 17 kB 00:00 2026-03-09T17:44:40.618 INFO:teuthology.orchestra.run.vm05.stdout:(33/143): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.6 MB/s | 299 kB 00:00 2026-03-09T17:44:40.725 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T17:44:40.780 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T17:44:40.780 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T17:44:40.843 INFO:teuthology.orchestra.run.vm05.stdout:(34/143): cephadm-19.2.3-678.ge911bdeb.el9.noar 3.3 MB/s | 769 kB 00:00 2026-03-09T17:44:41.097 INFO:teuthology.orchestra.run.vm05.stdout:(35/143): ceph-mgr-dashboard-19.2.3-678.ge911bd 2.2 MB/s | 3.8 MB 00:01 2026-03-09T17:44:41.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:40 vm01 ceph-mon[52793]: pgmap v144: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:41.230 INFO:teuthology.orchestra.run.vm05.stdout:(36/143): cryptsetup-2.8.1-3.el9.x86_64.rpm 908 kB/s | 351 kB 00:00 2026-03-09T17:44:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:40 vm04 ceph-mon[53158]: pgmap v144: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:41.304 INFO:teuthology.orchestra.run.vm05.stdout:(37/143): device-mapper-multipath-libs-0.8.7-45 3.8 MB/s | 289 kB 00:00 2026-03-09T17:44:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:40 vm05 ceph-mon[53831]: pgmap v144: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:41.377 INFO:teuthology.orchestra.run.vm05.stdout:(38/143): iscsi-initiator-utils-6.2.1.11-0.git4 5.3 MB/s | 392 kB 00:00 2026-03-09T17:44:41.417 INFO:teuthology.orchestra.run.vm05.stdout:(39/143): device-mapper-multipath-0.8.7-45.el9. 488 kB/s | 156 kB 00:00 2026-03-09T17:44:41.445 INFO:teuthology.orchestra.run.vm05.stdout:(40/143): iscsi-initiator-utils-iscsiuio-6.2.1. 1.2 MB/s | 81 kB 00:00 2026-03-09T17:44:41.486 INFO:teuthology.orchestra.run.vm05.stdout:(41/143): isns-utils-libs-0.101-4.el9.x86_64.rp 1.4 MB/s | 101 kB 00:00 2026-03-09T17:44:41.513 INFO:teuthology.orchestra.run.vm05.stdout:(42/143): ledmon-libs-1.1.0-3.el9.x86_64.rpm 600 kB/s | 40 kB 00:00 2026-03-09T17:44:41.554 INFO:teuthology.orchestra.run.vm05.stdout:(43/143): libconfig-1.7.2-9.el9.x86_64.rpm 1.0 MB/s | 72 kB 00:00 2026-03-09T17:44:41.592 INFO:teuthology.orchestra.run.vm05.stdout:(44/143): libgfortran-11.5.0-14.el9.x86_64.rpm 9.9 MB/s | 794 kB 00:00 2026-03-09T17:44:41.632 INFO:teuthology.orchestra.run.vm05.stdout:(45/143): libquadmath-11.5.0-14.el9.x86_64.rpm 2.3 MB/s | 184 kB 00:00 2026-03-09T17:44:41.638 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T17:44:41.638 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T17:44:41.656 INFO:teuthology.orchestra.run.vm05.stdout:(46/143): mailcap-2.1.49-5.el9.noarch.rpm 517 kB/s | 33 kB 00:00 2026-03-09T17:44:41.700 INFO:teuthology.orchestra.run.vm05.stdout:(47/143): pciutils-3.7.0-7.el9.x86_64.rpm 1.3 MB/s | 93 kB 00:00 2026-03-09T17:44:41.727 INFO:teuthology.orchestra.run.vm05.stdout:(48/143): python3-cffi-1.14.5-5.el9.x86_64.rpm 3.5 MB/s | 253 kB 00:00 2026-03-09T17:44:41.796 INFO:teuthology.orchestra.run.vm05.stdout:(49/143): python3-ply-3.11-14.el9.noarch.rpm 1.5 MB/s | 106 kB 00:00 2026-03-09T17:44:41.864 INFO:teuthology.orchestra.run.vm05.stdout:(50/143): python3-pycparser-2.20-6.el9.noarch.r 1.9 MB/s | 135 kB 00:00 2026-03-09T17:44:41.899 INFO:teuthology.orchestra.run.vm05.stdout:(51/143): python3-cryptography-36.0.1-5.el9.x86 6.3 MB/s | 1.2 MB 00:00 2026-03-09T17:44:41.933 INFO:teuthology.orchestra.run.vm05.stdout:(52/143): python3-pyparsing-2.4.7-9.el9.noarch. 2.2 MB/s | 150 kB 00:00 2026-03-09T17:44:41.968 INFO:teuthology.orchestra.run.vm05.stdout:(53/143): python3-requests-2.25.1-10.el9.noarch 1.8 MB/s | 126 kB 00:00 2026-03-09T17:44:42.003 INFO:teuthology.orchestra.run.vm05.stdout:(54/143): python3-urllib3-1.26.5-7.el9.noarch.r 3.0 MB/s | 218 kB 00:00 2026-03-09T17:44:42.037 INFO:teuthology.orchestra.run.vm05.stdout:(55/143): unzip-6.0-59.el9.x86_64.rpm 2.6 MB/s | 182 kB 00:00 2026-03-09T17:44:42.085 INFO:teuthology.orchestra.run.vm05.stdout:(56/143): zip-3.0-35.el9.x86_64.rpm 3.2 MB/s | 266 kB 00:00 2026-03-09T17:44:42.213 INFO:teuthology.orchestra.run.vm05.stdout:(57/143): ceph-test-19.2.3-678.ge911bdeb.el9.x8 6.9 MB/s | 50 MB 00:07 2026-03-09T17:44:42.390 INFO:teuthology.orchestra.run.vm05.stdout:(58/143): flexiblas-3.0.4-9.el9.x86_64.rpm 97 kB/s | 30 kB 00:00 2026-03-09T17:44:42.490 INFO:teuthology.orchestra.run.vm05.stdout:(59/143): flexiblas-openblas-openmp-3.0.4-9.el9 148 kB/s | 15 kB 00:00 2026-03-09T17:44:42.550 INFO:teuthology.orchestra.run.vm05.stdout:(60/143): boost-program-options-1.75.0-13.el9.x 203 kB/s | 104 kB 00:00 2026-03-09T17:44:42.559 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T17:44:42.574 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/146 2026-03-09T17:44:42.587 INFO:teuthology.orchestra.run.vm04.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/146 2026-03-09T17:44:42.724 INFO:teuthology.orchestra.run.vm05.stdout:(61/143): libpmemobj-1.12.1-1.el9.x86_64.rpm 926 kB/s | 160 kB 00:00 2026-03-09T17:44:42.761 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/146 2026-03-09T17:44:42.763 INFO:teuthology.orchestra.run.vm04.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-09T17:44:42.771 INFO:teuthology.orchestra.run.vm05.stdout:(62/143): libnbd-1.20.3-4.el9.x86_64.rpm 584 kB/s | 164 kB 00:00 2026-03-09T17:44:42.823 INFO:teuthology.orchestra.run.vm05.stdout:(63/143): librabbitmq-0.11.0-7.el9.x86_64.rpm 458 kB/s | 45 kB 00:00 2026-03-09T17:44:42.825 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-09T17:44:42.827 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/146 2026-03-09T17:44:42.857 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/146 2026-03-09T17:44:42.867 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/146 2026-03-09T17:44:42.871 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/146 2026-03-09T17:44:42.874 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/146 2026-03-09T17:44:42.879 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/146 2026-03-09T17:44:42.888 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/146 2026-03-09T17:44:42.890 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-09T17:44:42.927 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-09T17:44:42.929 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/146 2026-03-09T17:44:42.944 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/146 2026-03-09T17:44:42.947 INFO:teuthology.orchestra.run.vm05.stdout:(64/143): flexiblas-netlib-3.0.4-9.el9.x86_64.r 4.1 MB/s | 3.0 MB 00:00 2026-03-09T17:44:42.949 INFO:teuthology.orchestra.run.vm05.stdout:(65/143): libstoragemgmt-1.10.1-1.el9.x86_64.rp 1.9 MB/s | 246 kB 00:00 2026-03-09T17:44:42.980 INFO:teuthology.orchestra.run.vm04.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/146 2026-03-09T17:44:43.018 INFO:teuthology.orchestra.run.vm05.stdout:(66/143): librdkafka-1.6.1-102.el9.x86_64.rpm 2.6 MB/s | 662 kB 00:00 2026-03-09T17:44:43.019 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/146 2026-03-09T17:44:43.025 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/146 2026-03-09T17:44:43.048 INFO:teuthology.orchestra.run.vm05.stdout:(67/143): libxslt-1.1.34-12.el9.x86_64.rpm 2.3 MB/s | 233 kB 00:00 2026-03-09T17:44:43.051 INFO:teuthology.orchestra.run.vm04.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/146 2026-03-09T17:44:43.060 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/146 2026-03-09T17:44:43.070 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 18/146 2026-03-09T17:44:43.077 INFO:teuthology.orchestra.run.vm04.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 19/146 2026-03-09T17:44:43.081 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lua-5.4.4-4.el9.x86_64 20/146 2026-03-09T17:44:43.086 INFO:teuthology.orchestra.run.vm05.stdout:(68/143): lttng-ust-2.12.0-6.el9.x86_64.rpm 2.1 MB/s | 292 kB 00:00 2026-03-09T17:44:43.087 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 21/146 2026-03-09T17:44:43.116 INFO:teuthology.orchestra.run.vm04.stdout: Installing : unzip-6.0-59.el9.x86_64 22/146 2026-03-09T17:44:43.131 INFO:teuthology.orchestra.run.vm05.stdout:(69/143): lua-5.4.4-4.el9.x86_64.rpm 1.6 MB/s | 188 kB 00:00 2026-03-09T17:44:43.133 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 23/146 2026-03-09T17:44:43.137 INFO:teuthology.orchestra.run.vm05.stdout:(70/143): openblas-0.3.29-1.el9.x86_64.rpm 474 kB/s | 42 kB 00:00 2026-03-09T17:44:43.139 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 24/146 2026-03-09T17:44:43.147 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 25/146 2026-03-09T17:44:43.150 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 26/146 2026-03-09T17:44:43.184 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 27/146 2026-03-09T17:44:43.191 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 28/146 2026-03-09T17:44:43.202 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 29/146 2026-03-09T17:44:43.216 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 30/146 2026-03-09T17:44:43.225 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 31/146 2026-03-09T17:44:43.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:42 vm01 ceph-mon[52793]: pgmap v145: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:43.255 INFO:teuthology.orchestra.run.vm04.stdout: Installing : zip-3.0-35.el9.x86_64 32/146 2026-03-09T17:44:43.261 INFO:teuthology.orchestra.run.vm04.stdout: Installing : luarocks-3.9.2-5.el9.noarch 33/146 2026-03-09T17:44:43.270 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 34/146 2026-03-09T17:44:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:42 vm04 ceph-mon[53158]: pgmap v145: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:43.295 INFO:teuthology.orchestra.run.vm05.stdout:(71/143): protobuf-3.14.0-17.el9.x86_64.rpm 6.1 MB/s | 1.0 MB 00:00 2026-03-09T17:44:43.302 INFO:teuthology.orchestra.run.vm04.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 35/146 2026-03-09T17:44:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:42 vm05 ceph-mon[53831]: pgmap v145: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:43.364 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 36/146 2026-03-09T17:44:43.382 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 37/146 2026-03-09T17:44:43.385 INFO:teuthology.orchestra.run.vm05.stdout:(72/143): python3-devel-3.9.25-3.el9.x86_64.rpm 2.7 MB/s | 244 kB 00:00 2026-03-09T17:44:43.391 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rsa-4.9-2.el9.noarch 38/146 2026-03-09T17:44:43.401 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 39/146 2026-03-09T17:44:43.405 INFO:teuthology.orchestra.run.vm05.stdout:(73/143): python3-babel-2.9.1-2.el9.noarch.rpm 22 MB/s | 6.0 MB 00:00 2026-03-09T17:44:43.409 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 40/146 2026-03-09T17:44:43.416 INFO:teuthology.orchestra.run.vm04.stdout: Upgrading : kpartx-0.8.7-45.el9.x86_64 41/146 2026-03-09T17:44:43.420 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/146 2026-03-09T17:44:43.439 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/146 2026-03-09T17:44:43.466 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/146 2026-03-09T17:44:43.473 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/146 2026-03-09T17:44:43.475 INFO:teuthology.orchestra.run.vm05.stdout:(74/143): python3-jinja2-2.11.3-8.el9.noarch.rp 2.7 MB/s | 249 kB 00:00 2026-03-09T17:44:43.481 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/146 2026-03-09T17:44:43.496 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/146 2026-03-09T17:44:43.502 INFO:teuthology.orchestra.run.vm05.stdout:(75/143): python3-jmespath-1.0.1-1.el9.noarch.r 494 kB/s | 48 kB 00:00 2026-03-09T17:44:43.508 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/146 2026-03-09T17:44:43.519 INFO:teuthology.orchestra.run.vm05.stdout:(76/143): openblas-openmp-0.3.29-1.el9.x86_64.r 12 MB/s | 5.3 MB 00:00 2026-03-09T17:44:43.529 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/146 2026-03-09T17:44:43.558 INFO:teuthology.orchestra.run.vm05.stdout:(77/143): python3-libstoragemgmt-1.10.1-1.el9.x 2.1 MB/s | 177 kB 00:00 2026-03-09T17:44:43.603 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/146 2026-03-09T17:44:43.614 INFO:teuthology.orchestra.run.vm05.stdout:(78/143): python3-markupsafe-1.1.1-12.el9.x86_6 367 kB/s | 35 kB 00:00 2026-03-09T17:44:43.617 INFO:teuthology.orchestra.run.vm05.stdout:(79/143): python3-mako-1.1.4-6.el9.noarch.rpm 1.5 MB/s | 172 kB 00:00 2026-03-09T17:44:43.620 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/146 2026-03-09T17:44:43.632 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/146 2026-03-09T17:44:43.681 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/146 2026-03-09T17:44:43.704 INFO:teuthology.orchestra.run.vm05.stdout:(80/143): python3-packaging-20.9-5.el9.noarch.r 888 kB/s | 77 kB 00:00 2026-03-09T17:44:43.710 INFO:teuthology.orchestra.run.vm05.stdout:(81/143): python3-numpy-f2py-1.23.5-2.el9.x86_6 4.5 MB/s | 442 kB 00:00 2026-03-09T17:44:43.792 INFO:teuthology.orchestra.run.vm05.stdout:(82/143): python3-pyasn1-0.4.8-7.el9.noarch.rpm 1.9 MB/s | 157 kB 00:00 2026-03-09T17:44:43.802 INFO:teuthology.orchestra.run.vm05.stdout:(83/143): python3-protobuf-3.14.0-17.el9.noarch 2.7 MB/s | 267 kB 00:00 2026-03-09T17:44:43.910 INFO:teuthology.orchestra.run.vm05.stdout:(84/143): python3-requests-oauthlib-1.3.0-12.el 499 kB/s | 54 kB 00:00 2026-03-09T17:44:43.971 INFO:teuthology.orchestra.run.vm05.stdout:(85/143): python3-numpy-1.23.5-2.el9.x86_64.rpm 15 MB/s | 6.1 MB 00:00 2026-03-09T17:44:43.979 INFO:teuthology.orchestra.run.vm05.stdout:(86/143): python3-pyasn1-modules-0.4.8-7.el9.no 1.5 MB/s | 277 kB 00:00 2026-03-09T17:44:44.075 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/146 2026-03-09T17:44:44.111 INFO:teuthology.orchestra.run.vm05.stdout:(87/143): python3-toml-0.10.2-6.el9.noarch.rpm 298 kB/s | 42 kB 00:00 2026-03-09T17:44:44.126 INFO:teuthology.orchestra.run.vm05.stdout:(88/143): qatlib-25.08.0-2.el9.x86_64.rpm 1.6 MB/s | 240 kB 00:00 2026-03-09T17:44:44.162 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/146 2026-03-09T17:44:44.206 INFO:teuthology.orchestra.run.vm05.stdout:(89/143): qatlib-service-25.08.0-2.el9.x86_64.r 393 kB/s | 37 kB 00:00 2026-03-09T17:44:44.226 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/146 2026-03-09T17:44:44.282 INFO:teuthology.orchestra.run.vm05.stdout:(90/143): qatzip-libs-1.3.1-1.el9.x86_64.rpm 428 kB/s | 66 kB 00:00 2026-03-09T17:44:44.304 INFO:teuthology.orchestra.run.vm05.stdout:(91/143): socat-1.7.4.1-8.el9.x86_64.rpm 3.0 MB/s | 303 kB 00:00 2026-03-09T17:44:44.315 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/146 2026-03-09T17:44:44.405 INFO:teuthology.orchestra.run.vm05.stdout:(92/143): xmlstarlet-1.6.1-20.el9.x86_64.rpm 517 kB/s | 64 kB 00:00 2026-03-09T17:44:44.464 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/146 2026-03-09T17:44:44.514 INFO:teuthology.orchestra.run.vm05.stdout:(93/143): lua-devel-5.4.4-4.el9.x86_64.rpm 106 kB/s | 22 kB 00:00 2026-03-09T17:44:44.565 INFO:teuthology.orchestra.run.vm05.stdout:(94/143): abseil-cpp-20211102.0-4.el9.x86_64.rp 11 MB/s | 551 kB 00:00 2026-03-09T17:44:44.574 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/146 2026-03-09T17:44:44.581 INFO:teuthology.orchestra.run.vm05.stdout:(95/143): gperftools-libs-2.9.1-3.el9.x86_64.rp 19 MB/s | 308 kB 00:00 2026-03-09T17:44:44.591 INFO:teuthology.orchestra.run.vm05.stdout:(96/143): grpc-data-1.46.7-10.el9.noarch.rpm 2.0 MB/s | 19 kB 00:00 2026-03-09T17:44:44.595 INFO:teuthology.orchestra.run.vm04.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/146 2026-03-09T17:44:44.598 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/146 2026-03-09T17:44:44.640 INFO:teuthology.orchestra.run.vm04.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/146 2026-03-09T17:44:44.643 INFO:teuthology.orchestra.run.vm05.stdout:(97/143): protobuf-compiler-3.14.0-17.el9.x86_6 3.6 MB/s | 862 kB 00:00 2026-03-09T17:44:44.654 INFO:teuthology.orchestra.run.vm05.stdout:(98/143): libarrow-doc-9.0.0-15.el9.noarch.rpm 2.4 MB/s | 25 kB 00:00 2026-03-09T17:44:44.669 INFO:teuthology.orchestra.run.vm05.stdout:(99/143): libarrow-9.0.0-15.el9.x86_64.rpm 57 MB/s | 4.4 MB 00:00 2026-03-09T17:44:44.671 INFO:teuthology.orchestra.run.vm05.stdout:(100/143): liboath-2.6.12-1.el9.x86_64.rpm 2.8 MB/s | 49 kB 00:00 2026-03-09T17:44:44.672 INFO:teuthology.orchestra.run.vm05.stdout:(101/143): libunwind-1.6.2-1.el9.x86_64.rpm 24 MB/s | 67 kB 00:00 2026-03-09T17:44:44.678 INFO:teuthology.orchestra.run.vm05.stdout:(102/143): luarocks-3.9.2-5.el9.noarch.rpm 22 MB/s | 151 kB 00:00 2026-03-09T17:44:44.694 INFO:teuthology.orchestra.run.vm05.stdout:(103/143): python3-asyncssh-2.13.2-5.el9.noarch 35 MB/s | 548 kB 00:00 2026-03-09T17:44:44.699 INFO:teuthology.orchestra.run.vm05.stdout:(104/143): parquet-libs-9.0.0-15.el9.x86_64.rpm 31 MB/s | 838 kB 00:00 2026-03-09T17:44:44.699 INFO:teuthology.orchestra.run.vm05.stdout:(105/143): python3-autocommand-2.2.2-8.el9.noar 5.2 MB/s | 29 kB 00:00 2026-03-09T17:44:44.702 INFO:teuthology.orchestra.run.vm05.stdout:(106/143): python3-backports-tarfile-1.2.0-1.el 19 MB/s | 60 kB 00:00 2026-03-09T17:44:44.703 INFO:teuthology.orchestra.run.vm05.stdout:(107/143): python3-bcrypt-3.2.2-1.el9.x86_64.rp 13 MB/s | 43 kB 00:00 2026-03-09T17:44:44.706 INFO:teuthology.orchestra.run.vm05.stdout:(108/143): python3-cachetools-4.2.4-1.el9.noarc 9.7 MB/s | 32 kB 00:00 2026-03-09T17:44:44.706 INFO:teuthology.orchestra.run.vm05.stdout:(109/143): python3-certifi-2023.05.07-4.el9.noa 4.4 MB/s | 14 kB 00:00 2026-03-09T17:44:44.710 INFO:teuthology.orchestra.run.vm05.stdout:(110/143): python3-cheroot-10.0.1-4.el9.noarch. 40 MB/s | 173 kB 00:00 2026-03-09T17:44:44.710 INFO:teuthology.orchestra.run.vm04.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/146 2026-03-09T17:44:44.716 INFO:teuthology.orchestra.run.vm05.stdout:(111/143): python3-cherrypy-18.6.1-2.el9.noarch 37 MB/s | 358 kB 00:00 2026-03-09T17:44:44.719 INFO:teuthology.orchestra.run.vm05.stdout:(112/143): python3-google-auth-2.45.0-1.el9.noa 27 MB/s | 254 kB 00:00 2026-03-09T17:44:44.726 INFO:teuthology.orchestra.run.vm05.stdout:(113/143): python3-grpcio-tools-1.46.7-10.el9.x 22 MB/s | 144 kB 00:00 2026-03-09T17:44:44.730 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/146 2026-03-09T17:44:44.731 INFO:teuthology.orchestra.run.vm05.stdout:(114/143): python3-jaraco-8.2.1-3.el9.noarch.rp 2.4 MB/s | 11 kB 00:00 2026-03-09T17:44:44.735 INFO:teuthology.orchestra.run.vm05.stdout:(115/143): python3-jaraco-classes-3.2.1-5.el9.n 4.2 MB/s | 18 kB 00:00 2026-03-09T17:44:44.739 INFO:teuthology.orchestra.run.vm05.stdout:(116/143): python3-jaraco-collections-3.0.0-8.e 6.8 MB/s | 23 kB 00:00 2026-03-09T17:44:44.741 INFO:teuthology.orchestra.run.vm04.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/146 2026-03-09T17:44:44.747 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/146 2026-03-09T17:44:44.749 INFO:teuthology.orchestra.run.vm05.stdout:(117/143): python3-grpcio-1.46.7-10.el9.x86_64. 62 MB/s | 2.0 MB 00:00 2026-03-09T17:44:44.749 INFO:teuthology.orchestra.run.vm05.stdout:(118/143): python3-jaraco-context-6.0.1-3.el9.n 1.8 MB/s | 20 kB 00:00 2026-03-09T17:44:44.751 INFO:teuthology.orchestra.run.vm05.stdout:(119/143): python3-jaraco-functools-3.5.0-2.el9 8.2 MB/s | 19 kB 00:00 2026-03-09T17:44:44.752 INFO:teuthology.orchestra.run.vm05.stdout:(120/143): python3-jaraco-text-4.0.0-2.el9.noar 11 MB/s | 26 kB 00:00 2026-03-09T17:44:44.756 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/146 2026-03-09T17:44:44.757 INFO:teuthology.orchestra.run.vm05.stdout:(121/143): python3-logutils-0.3.5-21.el9.noarch 8.8 MB/s | 46 kB 00:00 2026-03-09T17:44:44.763 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/146 2026-03-09T17:44:44.764 INFO:teuthology.orchestra.run.vm05.stdout:(122/143): python3-more-itertools-8.12.0-2.el9. 11 MB/s | 79 kB 00:00 2026-03-09T17:44:44.769 INFO:teuthology.orchestra.run.vm05.stdout:(123/143): python3-natsort-7.1.1-5.el9.noarch.r 13 MB/s | 58 kB 00:00 2026-03-09T17:44:44.773 INFO:teuthology.orchestra.run.vm05.stdout:(124/143): python3-kubernetes-26.1.0-3.el9.noar 48 MB/s | 1.0 MB 00:00 2026-03-09T17:44:44.773 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/146 2026-03-09T17:44:44.776 INFO:teuthology.orchestra.run.vm05.stdout:(125/143): python3-pecan-1.4.2-3.el9.noarch.rpm 36 MB/s | 272 kB 00:00 2026-03-09T17:44:44.777 INFO:teuthology.orchestra.run.vm05.stdout:(126/143): python3-portend-3.1.0-2.el9.noarch.r 4.0 MB/s | 16 kB 00:00 2026-03-09T17:44:44.780 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/146 2026-03-09T17:44:44.780 INFO:teuthology.orchestra.run.vm05.stdout:(127/143): python3-pyOpenSSL-21.0.0-1.el9.noarc 24 MB/s | 90 kB 00:00 2026-03-09T17:44:44.781 INFO:teuthology.orchestra.run.vm05.stdout:(128/143): python3-repoze-lru-0.7-16.el9.noarch 8.4 MB/s | 31 kB 00:00 2026-03-09T17:44:44.785 INFO:teuthology.orchestra.run.vm05.stdout:(129/143): python3-routes-2.5.1-5.el9.noarch.rp 40 MB/s | 188 kB 00:00 2026-03-09T17:44:44.786 INFO:teuthology.orchestra.run.vm05.stdout:(130/143): python3-rsa-4.9-2.el9.noarch.rpm 10 MB/s | 59 kB 00:00 2026-03-09T17:44:44.787 INFO:teuthology.orchestra.run.vm05.stdout:(131/143): python3-tempora-5.0.0-2.el9.noarch.r 17 MB/s | 36 kB 00:00 2026-03-09T17:44:44.791 INFO:teuthology.orchestra.run.vm05.stdout:(132/143): python3-typing-extensions-4.15.0-1.e 21 MB/s | 86 kB 00:00 2026-03-09T17:44:44.793 INFO:teuthology.orchestra.run.vm05.stdout:(133/143): python3-webob-1.8.8-2.el9.noarch.rpm 41 MB/s | 230 kB 00:00 2026-03-09T17:44:44.795 INFO:teuthology.orchestra.run.vm05.stdout:(134/143): python3-websocket-client-1.2.3-2.el9 23 MB/s | 90 kB 00:00 2026-03-09T17:44:44.798 INFO:teuthology.orchestra.run.vm05.stdout:(135/143): python3-xmltodict-0.12.0-15.el9.noar 6.7 MB/s | 22 kB 00:00 2026-03-09T17:44:44.801 INFO:teuthology.orchestra.run.vm05.stdout:(136/143): python3-werkzeug-2.0.3-3.el9.1.noarc 57 MB/s | 427 kB 00:00 2026-03-09T17:44:44.801 INFO:teuthology.orchestra.run.vm05.stdout:(137/143): python3-zc-lockfile-2.0-10.el9.noarc 6.5 MB/s | 20 kB 00:00 2026-03-09T17:44:44.806 INFO:teuthology.orchestra.run.vm05.stdout:(138/143): re2-20211101-20.el9.x86_64.rpm 41 MB/s | 191 kB 00:00 2026-03-09T17:44:44.817 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/146 2026-03-09T17:44:44.826 INFO:teuthology.orchestra.run.vm05.stdout:(139/143): thrift-0.15.0-4.el9.x86_64.rpm 66 MB/s | 1.6 MB 00:00 2026-03-09T17:44:44.831 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/146 2026-03-09T17:44:44.881 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/146 2026-03-09T17:44:44.894 INFO:teuthology.orchestra.run.vm05.stdout:(140/143): python3-scipy-1.9.3-2.el9.x86_64.rpm 20 MB/s | 19 MB 00:00 2026-03-09T17:44:44.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:44 vm01 ceph-mon[52793]: pgmap v146: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:44.979 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:44:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:44:45.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:44 vm04 ceph-mon[53158]: pgmap v146: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:45.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:44 vm05 ceph-mon[53831]: pgmap v146: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:45.087 INFO:teuthology.orchestra.run.vm05.stdout:(141/143): kpartx-0.8.7-45.el9.x86_64.rpm 255 kB/s | 49 kB 00:00 2026-03-09T17:44:45.178 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 110/146 2026-03-09T17:44:45.179 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /sys 2026-03-09T17:44:45.179 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /proc 2026-03-09T17:44:45.179 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /mnt 2026-03-09T17:44:45.179 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /var/tmp 2026-03-09T17:44:45.179 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /home 2026-03-09T17:44:45.179 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /root 2026-03-09T17:44:45.179 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /tmp 2026-03-09T17:44:45.179 INFO:teuthology.orchestra.run.vm01.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc-sidecar@iscsi.iscsi.a:tcmu.service-pid 2026-03-09T17:44:45.179 INFO:teuthology.orchestra.run.vm01.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc@iscsi.iscsi.a.service-pid 2026-03-09T17:44:45.179 INFO:teuthology.orchestra.run.vm01.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc@mgr.x.service-pid 2026-03-09T17:44:45.179 INFO:teuthology.orchestra.run.vm01.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.a.service-pid 2026-03-09T17:44:45.179 INFO:teuthology.orchestra.run.vm01.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.0.service-pid 2026-03-09T17:44:45.179 INFO:teuthology.orchestra.run.vm01.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.1.service-pid 2026-03-09T17:44:45.179 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:45.188 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/146 2026-03-09T17:44:45.222 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/146 2026-03-09T17:44:45.229 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/146 2026-03-09T17:44:45.292 INFO:teuthology.orchestra.run.vm04.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/146 2026-03-09T17:44:45.296 INFO:teuthology.orchestra.run.vm04.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/146 2026-03-09T17:44:45.304 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 111/146 2026-03-09T17:44:45.323 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/146 2026-03-09T17:44:45.330 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 111/146 2026-03-09T17:44:45.330 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:44:45.330 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T17:44:45.330 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T17:44:45.331 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T17:44:45.331 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:45.565 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 112/146 2026-03-09T17:44:45.587 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 112/146 2026-03-09T17:44:45.588 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:44:45.588 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T17:44:45.588 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T17:44:45.588 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T17:44:45.588 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:45.596 INFO:teuthology.orchestra.run.vm01.stdout: Installing : mailcap-2.1.49-5.el9.noarch 113/146 2026-03-09T17:44:45.599 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 114/146 2026-03-09T17:44:45.618 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-09T17:44:45.618 INFO:teuthology.orchestra.run.vm01.stdout:Creating group 'qat' with GID 994. 2026-03-09T17:44:45.618 INFO:teuthology.orchestra.run.vm01.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T17:44:45.618 INFO:teuthology.orchestra.run.vm01.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T17:44:45.618 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:45.628 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-09T17:44:45.656 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-09T17:44:45.656 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T17:44:45.656 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:45.675 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 116/146 2026-03-09T17:44:45.683 INFO:teuthology.orchestra.run.vm01.stdout: Installing : isns-utils-libs-0.101-4.el9.x86_64 117/146 2026-03-09T17:44:45.686 INFO:teuthology.orchestra.run.vm01.stdout: Installing : iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 118/146 2026-03-09T17:44:45.702 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 118/146 2026-03-09T17:44:45.702 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/sockets.target.wants/iscsiuio.socket → /usr/lib/systemd/system/iscsiuio.socket. 2026-03-09T17:44:45.702 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:45.716 INFO:teuthology.orchestra.run.vm01.stdout: Installing : iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 119/146 2026-03-09T17:44:45.730 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/146 2026-03-09T17:44:45.733 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 119/146 2026-03-09T17:44:45.734 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/iscsi-starter.service → /usr/lib/systemd/system/iscsi-starter.service. 2026-03-09T17:44:45.734 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/sockets.target.wants/iscsid.socket → /usr/lib/systemd/system/iscsid.socket. 2026-03-09T17:44:45.734 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/iscsi-onboot.service → /usr/lib/systemd/system/iscsi-onboot.service. 2026-03-09T17:44:45.734 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:45.776 INFO:teuthology.orchestra.run.vm01.stdout: Installing : device-mapper-multipath-libs-0.8.7-45.el9.x86_64 120/146 2026-03-09T17:44:45.804 INFO:teuthology.orchestra.run.vm05.stdout:(142/143): librbd1-19.2.3-678.ge911bdeb.el9.x86 3.2 MB/s | 3.2 MB 00:00 2026-03-09T17:44:45.826 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/146 2026-03-09T17:44:45.853 INFO:teuthology.orchestra.run.vm01.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 121/146 2026-03-09T17:44:45.859 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 122/146 2026-03-09T17:44:45.873 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 122/146 2026-03-09T17:44:45.874 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:44:45.874 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T17:44:45.874 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:45.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:45 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:46.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:45 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:46.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:45 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:46.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:44:45 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:44:46.663 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 123/146 2026-03-09T17:44:46.667 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/146 2026-03-09T17:44:46.690 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 123/146 2026-03-09T17:44:46.691 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:44:46.691 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T17:44:46.691 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T17:44:46.691 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T17:44:46.691 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:46.704 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/146 2026-03-09T17:44:46.711 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/146 2026-03-09T17:44:46.717 INFO:teuthology.orchestra.run.vm04.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/146 2026-03-09T17:44:46.736 INFO:teuthology.orchestra.run.vm05.stdout:(143/143): librados2-19.2.3-678.ge911bdeb.el9.x 1.8 MB/s | 3.4 MB 00:01 2026-03-09T17:44:46.739 INFO:teuthology.orchestra.run.vm05.stdout:-------------------------------------------------------------------------------- 2026-03-09T17:44:46.740 INFO:teuthology.orchestra.run.vm05.stdout:Total 14 MB/s | 212 MB 00:14 2026-03-09T17:44:46.755 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 124/146 2026-03-09T17:44:46.758 INFO:teuthology.orchestra.run.vm01.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 124/146 2026-03-09T17:44:46.765 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 125/146 2026-03-09T17:44:46.792 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 126/146 2026-03-09T17:44:46.796 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 127/146 2026-03-09T17:44:46.886 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/146 2026-03-09T17:44:46.889 INFO:teuthology.orchestra.run.vm04.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/146 2026-03-09T17:44:46.924 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/146 2026-03-09T17:44:46.930 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/146 2026-03-09T17:44:46.939 INFO:teuthology.orchestra.run.vm04.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/146 2026-03-09T17:44:46.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:46 vm01 ceph-mon[52793]: pgmap v147: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:46.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:46 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:47.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:46 vm04 ceph-mon[53158]: pgmap v147: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:47.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:46 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:47.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:46 vm05 ceph-mon[53831]: pgmap v147: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:47.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:46 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:47.212 INFO:teuthology.orchestra.run.vm04.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/146 2026-03-09T17:44:47.215 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/146 2026-03-09T17:44:47.237 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/146 2026-03-09T17:44:47.239 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/146 2026-03-09T17:44:47.389 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 127/146 2026-03-09T17:44:47.397 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 128/146 2026-03-09T17:44:47.419 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-09T17:44:47.471 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-09T17:44:47.471 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-09T17:44:47.937 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 128/146 2026-03-09T17:44:47.940 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 129/146 2026-03-09T17:44:48.003 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 129/146 2026-03-09T17:44:48.062 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 130/146 2026-03-09T17:44:48.064 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 131/146 2026-03-09T17:44:48.086 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 131/146 2026-03-09T17:44:48.087 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:44:48.087 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T17:44:48.087 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T17:44:48.087 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T17:44:48.087 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:48.101 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 132/146 2026-03-09T17:44:48.113 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 132/146 2026-03-09T17:44:48.121 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 133/146 2026-03-09T17:44:48.124 INFO:teuthology.orchestra.run.vm01.stdout: Installing : device-mapper-multipath-0.8.7-45.el9.x86_64 134/146 2026-03-09T17:44:48.137 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: device-mapper-multipath-0.8.7-45.el9.x86_64 134/146 2026-03-09T17:44:48.138 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/multipathd.service → /usr/lib/systemd/system/multipathd.service. 2026-03-09T17:44:48.138 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/sockets.target.wants/multipathd.socket → /usr/lib/systemd/system/multipathd.socket. 2026-03-09T17:44:48.138 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:48.309 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-09T17:44:48.310 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-09T17:44:48.428 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-09T17:44:48.471 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-09T17:44:48.494 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-09T17:44:48.509 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 94/146 2026-03-09T17:44:48.519 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-packaging-20.9-5.el9.noarch 95/146 2026-03-09T17:44:48.538 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ply-3.11-14.el9.noarch 96/146 2026-03-09T17:44:48.561 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 97/146 2026-03-09T17:44:48.661 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 98/146 2026-03-09T17:44:48.677 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 99/146 2026-03-09T17:44:48.685 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 135/146 2026-03-09T17:44:48.713 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 135/146 2026-03-09T17:44:48.713 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:44:48.713 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T17:44:48.713 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T17:44:48.713 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T17:44:48.713 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:48.713 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 100/146 2026-03-09T17:44:48.723 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 136/146 2026-03-09T17:44:48.749 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 136/146 2026-03-09T17:44:48.749 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:44:48.749 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T17:44:48.749 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:48.753 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 101/146 2026-03-09T17:44:48.819 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 102/146 2026-03-09T17:44:48.829 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 103/146 2026-03-09T17:44:48.835 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 104/146 2026-03-09T17:44:48.843 INFO:teuthology.orchestra.run.vm04.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 105/146 2026-03-09T17:44:48.849 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 106/146 2026-03-09T17:44:48.851 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 107/146 2026-03-09T17:44:48.877 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 107/146 2026-03-09T17:44:48.915 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 137/146 2026-03-09T17:44:48.940 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 137/146 2026-03-09T17:44:48.940 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:44:48.940 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T17:44:48.940 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T17:44:48.940 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T17:44:48.940 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:48.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:48 vm01 ceph-mon[52793]: pgmap v148: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:49.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:48 vm04 ceph-mon[53158]: pgmap v148: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:49.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:48 vm05 ceph-mon[53831]: pgmap v148: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:49.225 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-09T17:44:49.241 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/146 2026-03-09T17:44:49.251 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 108/146 2026-03-09T17:44:49.255 INFO:teuthology.orchestra.run.vm05.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/146 2026-03-09T17:44:49.259 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 109/146 2026-03-09T17:44:49.306 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 109/146 2026-03-09T17:44:49.307 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T17:44:49.307 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:44:49.311 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 110/146 2026-03-09T17:44:49.433 INFO:teuthology.orchestra.run.vm05.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/146 2026-03-09T17:44:49.436 INFO:teuthology.orchestra.run.vm05.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-09T17:44:49.499 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-09T17:44:49.500 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/146 2026-03-09T17:44:49.532 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/146 2026-03-09T17:44:49.543 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/146 2026-03-09T17:44:49.546 INFO:teuthology.orchestra.run.vm05.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/146 2026-03-09T17:44:49.549 INFO:teuthology.orchestra.run.vm05.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/146 2026-03-09T17:44:49.554 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/146 2026-03-09T17:44:49.565 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/146 2026-03-09T17:44:49.598 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-09T17:44:49.637 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-09T17:44:49.639 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/146 2026-03-09T17:44:49.655 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/146 2026-03-09T17:44:49.688 INFO:teuthology.orchestra.run.vm05.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/146 2026-03-09T17:44:49.730 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/146 2026-03-09T17:44:49.735 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/146 2026-03-09T17:44:49.760 INFO:teuthology.orchestra.run.vm05.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/146 2026-03-09T17:44:49.768 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/146 2026-03-09T17:44:49.778 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 18/146 2026-03-09T17:44:49.785 INFO:teuthology.orchestra.run.vm05.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 19/146 2026-03-09T17:44:49.789 INFO:teuthology.orchestra.run.vm05.stdout: Installing : lua-5.4.4-4.el9.x86_64 20/146 2026-03-09T17:44:49.794 INFO:teuthology.orchestra.run.vm05.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 21/146 2026-03-09T17:44:49.824 INFO:teuthology.orchestra.run.vm05.stdout: Installing : unzip-6.0-59.el9.x86_64 22/146 2026-03-09T17:44:49.840 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 23/146 2026-03-09T17:44:49.845 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 24/146 2026-03-09T17:44:49.853 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 25/146 2026-03-09T17:44:49.856 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 26/146 2026-03-09T17:44:49.888 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 27/146 2026-03-09T17:44:49.895 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 28/146 2026-03-09T17:44:49.906 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 29/146 2026-03-09T17:44:49.921 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 30/146 2026-03-09T17:44:49.929 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 31/146 2026-03-09T17:44:49.960 INFO:teuthology.orchestra.run.vm05.stdout: Installing : zip-3.0-35.el9.x86_64 32/146 2026-03-09T17:44:49.976 INFO:teuthology.orchestra.run.vm05.stdout: Installing : luarocks-3.9.2-5.el9.noarch 33/146 2026-03-09T17:44:49.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:49 vm01 ceph-mon[52793]: from='client.? 192.168.123.104:0/1730306550' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-09T17:44:49.984 INFO:teuthology.orchestra.run.vm05.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 34/146 2026-03-09T17:44:50.014 INFO:teuthology.orchestra.run.vm05.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 35/146 2026-03-09T17:44:50.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:49 vm04 ceph-mon[53158]: from='client.? 192.168.123.104:0/1730306550' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-09T17:44:50.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:49 vm05 ceph-mon[53831]: from='client.? 192.168.123.104:0/1730306550' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-09T17:44:50.075 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 36/146 2026-03-09T17:44:50.092 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 37/146 2026-03-09T17:44:50.099 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-rsa-4.9-2.el9.noarch 38/146 2026-03-09T17:44:50.108 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 39/146 2026-03-09T17:44:50.112 INFO:teuthology.orchestra.run.vm05.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 40/146 2026-03-09T17:44:50.118 INFO:teuthology.orchestra.run.vm05.stdout: Upgrading : kpartx-0.8.7-45.el9.x86_64 41/146 2026-03-09T17:44:50.122 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/146 2026-03-09T17:44:50.140 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/146 2026-03-09T17:44:50.165 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/146 2026-03-09T17:44:50.172 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/146 2026-03-09T17:44:50.178 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/146 2026-03-09T17:44:50.194 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/146 2026-03-09T17:44:50.207 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/146 2026-03-09T17:44:50.219 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/146 2026-03-09T17:44:50.288 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/146 2026-03-09T17:44:50.297 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/146 2026-03-09T17:44:50.306 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/146 2026-03-09T17:44:50.355 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/146 2026-03-09T17:44:50.732 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/146 2026-03-09T17:44:50.747 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/146 2026-03-09T17:44:50.753 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/146 2026-03-09T17:44:50.762 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/146 2026-03-09T17:44:50.768 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/146 2026-03-09T17:44:50.776 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/146 2026-03-09T17:44:50.780 INFO:teuthology.orchestra.run.vm05.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/146 2026-03-09T17:44:50.783 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/146 2026-03-09T17:44:50.819 INFO:teuthology.orchestra.run.vm05.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/146 2026-03-09T17:44:50.875 INFO:teuthology.orchestra.run.vm05.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/146 2026-03-09T17:44:50.889 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/146 2026-03-09T17:44:50.898 INFO:teuthology.orchestra.run.vm05.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/146 2026-03-09T17:44:50.903 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/146 2026-03-09T17:44:50.911 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/146 2026-03-09T17:44:50.917 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/146 2026-03-09T17:44:50.927 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/146 2026-03-09T17:44:50.932 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/146 2026-03-09T17:44:50.967 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/146 2026-03-09T17:44:50.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:50 vm01 ceph-mon[52793]: pgmap v149: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:50.981 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/146 2026-03-09T17:44:51.026 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/146 2026-03-09T17:44:51.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:50 vm04 ceph-mon[53158]: pgmap v149: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:51.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:50 vm05 ceph-mon[53831]: pgmap v149: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:51.307 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/146 2026-03-09T17:44:51.339 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/146 2026-03-09T17:44:51.345 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/146 2026-03-09T17:44:51.407 INFO:teuthology.orchestra.run.vm05.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/146 2026-03-09T17:44:51.410 INFO:teuthology.orchestra.run.vm05.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/146 2026-03-09T17:44:51.435 INFO:teuthology.orchestra.run.vm05.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/146 2026-03-09T17:44:51.823 INFO:teuthology.orchestra.run.vm05.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/146 2026-03-09T17:44:51.913 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/146 2026-03-09T17:44:51.918 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 138/146 2026-03-09T17:44:51.930 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 139/146 2026-03-09T17:44:51.938 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 140/146 2026-03-09T17:44:52.001 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 141/146 2026-03-09T17:44:52.012 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 142/146 2026-03-09T17:44:52.016 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 143/146 2026-03-09T17:44:52.016 INFO:teuthology.orchestra.run.vm01.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-09T17:44:52.034 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-09T17:44:52.034 INFO:teuthology.orchestra.run.vm01.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 145/146 2026-03-09T17:44:52.052 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 145/146 2026-03-09T17:44:52.072 INFO:teuthology.orchestra.run.vm01.stdout: Cleanup : kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-09T17:44:52.768 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/146 2026-03-09T17:44:52.911 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/146 2026-03-09T17:44:52.920 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/146 2026-03-09T17:44:52.925 INFO:teuthology.orchestra.run.vm05.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/146 2026-03-09T17:44:53.107 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/146 2026-03-09T17:44:53.110 INFO:teuthology.orchestra.run.vm05.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/146 2026-03-09T17:44:53.146 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/146 2026-03-09T17:44:53.223 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/146 2026-03-09T17:44:53.236 INFO:teuthology.orchestra.run.vm05.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/146 2026-03-09T17:44:53.418 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-09T17:44:53.418 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/146 2026-03-09T17:44:53.418 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/146 2026-03-09T17:44:53.418 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/146 2026-03-09T17:44:53.418 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-09T17:44:53.418 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/146 2026-03-09T17:44:53.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/146 2026-03-09T17:44:53.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/146 2026-03-09T17:44:53.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/146 2026-03-09T17:44:53.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/146 2026-03-09T17:44:53.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/146 2026-03-09T17:44:53.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/146 2026-03-09T17:44:53.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/146 2026-03-09T17:44:53.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/146 2026-03-09T17:44:53.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/146 2026-03-09T17:44:53.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/146 2026-03-09T17:44:53.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/146 2026-03-09T17:44:53.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/146 2026-03-09T17:44:53.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : device-mapper-multipath-0.8.7-45.el9.x86_64 38/146 2026-03-09T17:44:53.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : device-mapper-multipath-libs-0.8.7-45.el9.x86_64 39/146 2026-03-09T17:44:53.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 40/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 41/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : isns-utils-libs-0.101-4.el9.x86_64 42/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 43/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 44/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 45/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 46/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 47/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 48/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 49/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 50/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ply-3.11-14.el9.noarch 51/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 52/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 53/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 54/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 55/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : unzip-6.0-59.el9.x86_64 56/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : zip-3.0-35.el9.x86_64 57/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 58/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 59/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 60/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 61/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 62/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 63/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 64/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 65/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 66/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 67/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 68/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-5.4.4-4.el9.x86_64 69/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 70/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 71/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 72/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 73/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 74/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 75/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 76/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 77/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 78/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 79/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 80/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 81/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 82/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 83/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 84/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 85/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 87/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 88/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 89/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 90/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 91/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 92/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 93/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 94/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 95/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 96/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 97/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 98/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 99/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 100/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 101/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 102/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 103/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 104/146 2026-03-09T17:44:53.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 105/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 106/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 107/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 108/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 109/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 110/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 111/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 112/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 113/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 114/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 115/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 116/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 117/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 118/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 119/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 120/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 121/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 122/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 123/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 124/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 125/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 126/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 127/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 128/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 129/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 130/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 131/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 132/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 133/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 134/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 135/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 136/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 137/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 138/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : re2-1:20211101-20.el9.x86_64 139/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 140/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 141/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 142/146 2026-03-09T17:44:53.424 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 143/146 2026-03-09T17:44:53.425 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-09T17:44:53.425 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : kpartx-0.8.7-45.el9.x86_64 145/146 2026-03-09T17:44:53.532 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-09T17:44:53.532 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:53.532 INFO:teuthology.orchestra.run.vm01.stdout:Upgraded: 2026-03-09T17:44:53.532 INFO:teuthology.orchestra.run.vm01.stdout: kpartx-0.8.7-45.el9.x86_64 2026-03-09T17:44:53.532 INFO:teuthology.orchestra.run.vm01.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.532 INFO:teuthology.orchestra.run.vm01.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.532 INFO:teuthology.orchestra.run.vm01.stdout:Installed: 2026-03-09T17:44:53.532 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T17:44:53.532 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T17:44:53.532 INFO:teuthology.orchestra.run.vm01.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.532 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.532 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.532 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: device-mapper-multipath-0.8.7-45.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: device-mapper-multipath-libs-0.8.7-45.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3e853.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: isns-utils-libs-0.101-4.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T17:44:53.533 INFO:teuthology.orchestra.run.vm01.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T17:44:53.534 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyparsing-2.4.7-9.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T17:44:53.535 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: zip-3.0-35.el9.x86_64 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:44:53.536 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:44:53.565 INFO:teuthology.orchestra.run.vm05.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/146 2026-03-09T17:44:53.568 INFO:teuthology.orchestra.run.vm05.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/146 2026-03-09T17:44:53.600 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/146 2026-03-09T17:44:53.623 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/146 2026-03-09T17:44:53.641 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:53 vm05 ceph-mon[53831]: pgmap v150: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:53.642 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:53 vm01 ceph-mon[52793]: pgmap v150: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:53.645 DEBUG:teuthology.parallel:result is None 2026-03-09T17:44:53.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:53 vm04 ceph-mon[53158]: pgmap v150: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:54.671 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:54 vm01 ceph-mon[52793]: pgmap v151: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:54.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:54 vm04 ceph-mon[53158]: pgmap v151: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:54.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:54 vm05 ceph-mon[53831]: pgmap v151: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:54.912 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-09T17:44:54.930 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-09T17:44:54.961 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-09T17:44:54.977 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 94/146 2026-03-09T17:44:54.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:44:54 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:44:54.988 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-packaging-20.9-5.el9.noarch 95/146 2026-03-09T17:44:55.007 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-ply-3.11-14.el9.noarch 96/146 2026-03-09T17:44:55.033 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 97/146 2026-03-09T17:44:55.147 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 98/146 2026-03-09T17:44:55.164 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 99/146 2026-03-09T17:44:55.199 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 100/146 2026-03-09T17:44:55.245 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 101/146 2026-03-09T17:44:55.318 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 102/146 2026-03-09T17:44:55.453 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 103/146 2026-03-09T17:44:55.555 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 104/146 2026-03-09T17:44:55.620 INFO:teuthology.orchestra.run.vm05.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 105/146 2026-03-09T17:44:55.684 INFO:teuthology.orchestra.run.vm05.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 106/146 2026-03-09T17:44:55.696 INFO:teuthology.orchestra.run.vm05.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 107/146 2026-03-09T17:44:55.714 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 107/146 2026-03-09T17:44:56.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:56 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:56.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:44:55 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:44:56.075 INFO:teuthology.orchestra.run.vm05.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 108/146 2026-03-09T17:44:56.082 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 109/146 2026-03-09T17:44:56.138 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 109/146 2026-03-09T17:44:56.138 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T17:44:56.138 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:44:56.141 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 110/146 2026-03-09T17:44:56.292 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:56 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:56.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:56 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:56.610 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 110/146 2026-03-09T17:44:56.610 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /sys 2026-03-09T17:44:56.611 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /proc 2026-03-09T17:44:56.611 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /mnt 2026-03-09T17:44:56.611 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /var/tmp 2026-03-09T17:44:56.611 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /home 2026-03-09T17:44:56.611 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /root 2026-03-09T17:44:56.611 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /tmp 2026-03-09T17:44:56.611 INFO:teuthology.orchestra.run.vm04.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.b.service-pid 2026-03-09T17:44:56.611 INFO:teuthology.orchestra.run.vm04.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.2.service-pid 2026-03-09T17:44:56.611 INFO:teuthology.orchestra.run.vm04.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.3.service-pid 2026-03-09T17:44:56.611 INFO:teuthology.orchestra.run.vm04.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.4.service-pid 2026-03-09T17:44:56.611 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:44:56.740 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 111/146 2026-03-09T17:44:56.767 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 111/146 2026-03-09T17:44:56.767 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:44:56.767 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T17:44:56.767 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T17:44:56.767 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T17:44:56.767 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:44:57.006 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 112/146 2026-03-09T17:44:57.021 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:57 vm04 ceph-mon[53158]: pgmap v152: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:57.021 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:57 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:57.021 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:57 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3403605030' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-09T17:44:57.031 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 112/146 2026-03-09T17:44:57.031 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:44:57.031 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T17:44:57.031 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T17:44:57.031 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T17:44:57.031 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:44:57.041 INFO:teuthology.orchestra.run.vm04.stdout: Installing : mailcap-2.1.49-5.el9.noarch 113/146 2026-03-09T17:44:57.044 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 114/146 2026-03-09T17:44:57.063 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-09T17:44:57.063 INFO:teuthology.orchestra.run.vm04.stdout:Creating group 'qat' with GID 994. 2026-03-09T17:44:57.063 INFO:teuthology.orchestra.run.vm04.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T17:44:57.063 INFO:teuthology.orchestra.run.vm04.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T17:44:57.063 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:44:57.075 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-09T17:44:57.104 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-09T17:44:57.104 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T17:44:57.104 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:44:57.123 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 116/146 2026-03-09T17:44:57.128 INFO:teuthology.orchestra.run.vm04.stdout: Installing : isns-utils-libs-0.101-4.el9.x86_64 117/146 2026-03-09T17:44:57.130 INFO:teuthology.orchestra.run.vm04.stdout: Installing : iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 118/146 2026-03-09T17:44:57.146 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 118/146 2026-03-09T17:44:57.146 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/sockets.target.wants/iscsiuio.socket → /usr/lib/systemd/system/iscsiuio.socket. 2026-03-09T17:44:57.146 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:44:57.161 INFO:teuthology.orchestra.run.vm04.stdout: Installing : iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 119/146 2026-03-09T17:44:57.182 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 119/146 2026-03-09T17:44:57.182 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/iscsi-starter.service → /usr/lib/systemd/system/iscsi-starter.service. 2026-03-09T17:44:57.182 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/sockets.target.wants/iscsid.socket → /usr/lib/systemd/system/iscsid.socket. 2026-03-09T17:44:57.182 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/iscsi-onboot.service → /usr/lib/systemd/system/iscsi-onboot.service. 2026-03-09T17:44:57.182 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:44:57.230 INFO:teuthology.orchestra.run.vm04.stdout: Installing : device-mapper-multipath-libs-0.8.7-45.el9.x86_64 120/146 2026-03-09T17:44:57.311 INFO:teuthology.orchestra.run.vm04.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 121/146 2026-03-09T17:44:57.316 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 122/146 2026-03-09T17:44:57.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:57 vm05 ceph-mon[53831]: pgmap v152: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:57.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:57 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:57.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:57 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3403605030' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-09T17:44:57.334 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 122/146 2026-03-09T17:44:57.334 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:44:57.334 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T17:44:57.334 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:44:57.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:57 vm01 ceph-mon[52793]: pgmap v152: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:44:57.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:57 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:44:57.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:57 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3403605030' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-09T17:44:58.187 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 123/146 2026-03-09T17:44:58.217 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 123/146 2026-03-09T17:44:58.217 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:44:58.217 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T17:44:58.217 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T17:44:58.217 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T17:44:58.217 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:44:58.284 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 124/146 2026-03-09T17:44:58.288 INFO:teuthology.orchestra.run.vm04.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 124/146 2026-03-09T17:44:58.294 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 125/146 2026-03-09T17:44:58.318 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 126/146 2026-03-09T17:44:58.322 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 127/146 2026-03-09T17:44:58.917 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 127/146 2026-03-09T17:44:58.959 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 128/146 2026-03-09T17:44:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:44:59 vm04 ceph-mon[53158]: pgmap v153: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:59.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:44:59 vm05 ceph-mon[53831]: pgmap v153: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:59.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:44:59 vm01 ceph-mon[52793]: pgmap v153: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:44:59.538 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 128/146 2026-03-09T17:44:59.541 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 129/146 2026-03-09T17:44:59.613 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 129/146 2026-03-09T17:44:59.683 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 130/146 2026-03-09T17:44:59.686 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 131/146 2026-03-09T17:44:59.715 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 131/146 2026-03-09T17:44:59.715 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:44:59.715 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T17:44:59.715 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T17:44:59.715 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T17:44:59.715 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:44:59.729 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 132/146 2026-03-09T17:44:59.743 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 132/146 2026-03-09T17:44:59.752 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 133/146 2026-03-09T17:44:59.755 INFO:teuthology.orchestra.run.vm04.stdout: Installing : device-mapper-multipath-0.8.7-45.el9.x86_64 134/146 2026-03-09T17:44:59.773 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: device-mapper-multipath-0.8.7-45.el9.x86_64 134/146 2026-03-09T17:44:59.773 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/multipathd.service → /usr/lib/systemd/system/multipathd.service. 2026-03-09T17:44:59.773 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/sockets.target.wants/multipathd.socket → /usr/lib/systemd/system/multipathd.socket. 2026-03-09T17:44:59.773 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:45:00.365 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 135/146 2026-03-09T17:45:00.391 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 135/146 2026-03-09T17:45:00.391 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:45:00.391 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T17:45:00.391 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T17:45:00.391 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T17:45:00.391 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:45:00.403 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 136/146 2026-03-09T17:45:00.428 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 136/146 2026-03-09T17:45:00.428 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:45:00.428 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T17:45:00.428 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:45:00.609 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 137/146 2026-03-09T17:45:00.638 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 137/146 2026-03-09T17:45:00.638 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:45:00.638 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T17:45:00.638 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T17:45:00.638 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T17:45:00.638 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:45:01.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:01 vm01 ceph-mon[52793]: pgmap v154: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:01.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:01 vm04 ceph-mon[53158]: pgmap v154: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:01.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:01 vm05 ceph-mon[53831]: pgmap v154: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:02.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:02 vm01 ceph-mon[52793]: pgmap v155: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:03.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:02 vm04 ceph-mon[53158]: pgmap v155: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:03.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:02 vm05 ceph-mon[53831]: pgmap v155: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:03.191 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 110/146 2026-03-09T17:45:03.191 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /sys 2026-03-09T17:45:03.192 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /proc 2026-03-09T17:45:03.192 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /mnt 2026-03-09T17:45:03.192 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /var/tmp 2026-03-09T17:45:03.192 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /home 2026-03-09T17:45:03.192 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /root 2026-03-09T17:45:03.192 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /tmp 2026-03-09T17:45:03.192 INFO:teuthology.orchestra.run.vm05.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc-sidecar@iscsi.iscsi.b:tcmu.service-pid 2026-03-09T17:45:03.192 INFO:teuthology.orchestra.run.vm05.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc@iscsi.iscsi.b.service-pid 2026-03-09T17:45:03.192 INFO:teuthology.orchestra.run.vm05.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.c.service-pid 2026-03-09T17:45:03.192 INFO:teuthology.orchestra.run.vm05.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.5.service-pid 2026-03-09T17:45:03.192 INFO:teuthology.orchestra.run.vm05.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.6.service-pid 2026-03-09T17:45:03.192 INFO:teuthology.orchestra.run.vm05.stdout:Warning no default label for /run/ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.7.service-pid 2026-03-09T17:45:03.192 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:03.313 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 111/146 2026-03-09T17:45:03.342 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 111/146 2026-03-09T17:45:03.342 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:45:03.342 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T17:45:03.342 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T17:45:03.342 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T17:45:03.342 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:03.573 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 112/146 2026-03-09T17:45:03.598 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 112/146 2026-03-09T17:45:03.598 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:45:03.598 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T17:45:03.598 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T17:45:03.598 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T17:45:03.598 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:03.607 INFO:teuthology.orchestra.run.vm05.stdout: Installing : mailcap-2.1.49-5.el9.noarch 113/146 2026-03-09T17:45:03.610 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 114/146 2026-03-09T17:45:03.629 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 138/146 2026-03-09T17:45:03.631 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-09T17:45:03.631 INFO:teuthology.orchestra.run.vm05.stdout:Creating group 'qat' with GID 994. 2026-03-09T17:45:03.631 INFO:teuthology.orchestra.run.vm05.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T17:45:03.631 INFO:teuthology.orchestra.run.vm05.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T17:45:03.631 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:03.642 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 139/146 2026-03-09T17:45:03.642 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-09T17:45:03.648 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 140/146 2026-03-09T17:45:03.672 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-09T17:45:03.672 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T17:45:03.672 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:03.709 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 141/146 2026-03-09T17:45:03.728 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 116/146 2026-03-09T17:45:03.745 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 142/146 2026-03-09T17:45:03.751 INFO:teuthology.orchestra.run.vm05.stdout: Installing : isns-utils-libs-0.101-4.el9.x86_64 117/146 2026-03-09T17:45:03.809 INFO:teuthology.orchestra.run.vm05.stdout: Installing : iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 118/146 2026-03-09T17:45:03.811 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 143/146 2026-03-09T17:45:03.811 INFO:teuthology.orchestra.run.vm04.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-09T17:45:03.827 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 118/146 2026-03-09T17:45:03.827 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/sockets.target.wants/iscsiuio.socket → /usr/lib/systemd/system/iscsiuio.socket. 2026-03-09T17:45:03.827 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:03.898 INFO:teuthology.orchestra.run.vm05.stdout: Installing : iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 119/146 2026-03-09T17:45:03.915 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 119/146 2026-03-09T17:45:03.915 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/iscsi-starter.service → /usr/lib/systemd/system/iscsi-starter.service. 2026-03-09T17:45:03.915 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/sockets.target.wants/iscsid.socket → /usr/lib/systemd/system/iscsid.socket. 2026-03-09T17:45:03.915 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/iscsi-onboot.service → /usr/lib/systemd/system/iscsi-onboot.service. 2026-03-09T17:45:03.915 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:03.981 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-09T17:45:03.981 INFO:teuthology.orchestra.run.vm04.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 145/146 2026-03-09T17:45:04.020 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 145/146 2026-03-09T17:45:04.038 INFO:teuthology.orchestra.run.vm05.stdout: Installing : device-mapper-multipath-libs-0.8.7-45.el9.x86_64 120/146 2026-03-09T17:45:04.044 INFO:teuthology.orchestra.run.vm04.stdout: Cleanup : kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-09T17:45:04.117 INFO:teuthology.orchestra.run.vm05.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 121/146 2026-03-09T17:45:04.122 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 122/146 2026-03-09T17:45:04.137 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 122/146 2026-03-09T17:45:04.137 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:45:04.137 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T17:45:04.137 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:04.977 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:45:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:45:04.994 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 123/146 2026-03-09T17:45:05.020 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 123/146 2026-03-09T17:45:05.020 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:45:05.020 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T17:45:05.020 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T17:45:05.020 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T17:45:05.020 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:05.160 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 124/146 2026-03-09T17:45:05.182 INFO:teuthology.orchestra.run.vm05.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 124/146 2026-03-09T17:45:05.294 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 125/146 2026-03-09T17:45:05.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:05 vm05 ceph-mon[53831]: pgmap v156: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:05.437 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 126/146 2026-03-09T17:45:05.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:05 vm01 ceph-mon[52793]: pgmap v156: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:05.488 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 127/146 2026-03-09T17:45:05.563 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-09T17:45:05.563 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/146 2026-03-09T17:45:05.563 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/146 2026-03-09T17:45:05.563 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/146 2026-03-09T17:45:05.564 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-09T17:45:05.564 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/146 2026-03-09T17:45:05.564 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/146 2026-03-09T17:45:05.564 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/146 2026-03-09T17:45:05.564 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/146 2026-03-09T17:45:05.564 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/146 2026-03-09T17:45:05.564 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/146 2026-03-09T17:45:05.564 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-09T17:45:05.564 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/146 2026-03-09T17:45:05.564 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/146 2026-03-09T17:45:05.564 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/146 2026-03-09T17:45:05.564 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/146 2026-03-09T17:45:05.564 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : device-mapper-multipath-0.8.7-45.el9.x86_64 38/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : device-mapper-multipath-libs-0.8.7-45.el9.x86_64 39/146 2026-03-09T17:45:05.565 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 40/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 41/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : isns-utils-libs-0.101-4.el9.x86_64 42/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 43/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 44/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 45/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 46/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 47/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 48/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 49/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 50/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ply-3.11-14.el9.noarch 51/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 52/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 53/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 54/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 55/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : unzip-6.0-59.el9.x86_64 56/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : zip-3.0-35.el9.x86_64 57/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 58/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 59/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 60/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 61/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 62/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 63/146 2026-03-09T17:45:05.566 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 64/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 65/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 66/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 67/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 68/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-5.4.4-4.el9.x86_64 69/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 70/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 71/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 72/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 73/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 74/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 75/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 76/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 77/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 78/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 79/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 80/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 81/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 82/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 83/146 2026-03-09T17:45:05.567 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 84/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 85/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 87/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 88/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 89/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 90/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 91/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 92/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 93/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 94/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 95/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 96/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 97/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 98/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 99/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 100/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 101/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 102/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 103/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 104/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 105/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 106/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 107/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 108/146 2026-03-09T17:45:05.568 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 109/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 110/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 111/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 112/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 113/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 114/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 115/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 116/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 117/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 118/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 119/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 120/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 121/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 122/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 123/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 124/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 125/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 126/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 127/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 128/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 129/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 130/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 131/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 132/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 133/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 134/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 135/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 136/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 137/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 138/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : re2-1:20211101-20.el9.x86_64 139/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 140/146 2026-03-09T17:45:05.569 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 141/146 2026-03-09T17:45:05.570 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 142/146 2026-03-09T17:45:05.570 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 143/146 2026-03-09T17:45:05.570 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-09T17:45:05.570 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : kpartx-0.8.7-45.el9.x86_64 145/146 2026-03-09T17:45:05.685 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-09T17:45:05.685 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:45:05.685 INFO:teuthology.orchestra.run.vm04.stdout:Upgraded: 2026-03-09T17:45:05.685 INFO:teuthology.orchestra.run.vm04.stdout: kpartx-0.8.7-45.el9.x86_64 2026-03-09T17:45:05.685 INFO:teuthology.orchestra.run.vm04.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.685 INFO:teuthology.orchestra.run.vm04.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.685 INFO:teuthology.orchestra.run.vm04.stdout:Installed: 2026-03-09T17:45:05.685 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T17:45:05.685 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T17:45:05.685 INFO:teuthology.orchestra.run.vm04.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.685 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.685 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: device-mapper-multipath-0.8.7-45.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: device-mapper-multipath-libs-0.8.7-45.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3e853.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: isns-utils-libs-0.101-4.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T17:45:05.686 INFO:teuthology.orchestra.run.vm04.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyparsing-2.4.7-9.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T17:45:05.687 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: zip-3.0-35.el9.x86_64 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:45:05.688 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:45:05.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:05 vm04 ceph-mon[53158]: pgmap v156: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:05.798 DEBUG:teuthology.parallel:result is None 2026-03-09T17:45:06.074 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 127/146 2026-03-09T17:45:06.080 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 128/146 2026-03-09T17:45:06.187 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:45:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:45:06.187 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:06 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:06.187 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:06 vm05 ceph-mon[53831]: pgmap v157: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:06.187 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:06 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:06.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:06 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:06.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:06 vm01 ceph-mon[52793]: pgmap v157: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:06.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:06 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:06.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:06 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:06.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:06 vm04 ceph-mon[53158]: pgmap v157: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:06.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:06 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:06.618 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 128/146 2026-03-09T17:45:06.621 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 129/146 2026-03-09T17:45:06.684 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 129/146 2026-03-09T17:45:06.747 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 130/146 2026-03-09T17:45:06.799 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 131/146 2026-03-09T17:45:06.823 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 131/146 2026-03-09T17:45:06.823 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:45:06.823 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T17:45:06.823 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T17:45:06.823 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T17:45:06.823 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:06.965 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 132/146 2026-03-09T17:45:06.977 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 132/146 2026-03-09T17:45:07.090 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 133/146 2026-03-09T17:45:07.122 INFO:teuthology.orchestra.run.vm05.stdout: Installing : device-mapper-multipath-0.8.7-45.el9.x86_64 134/146 2026-03-09T17:45:07.134 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: device-mapper-multipath-0.8.7-45.el9.x86_64 134/146 2026-03-09T17:45:07.135 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/multipathd.service → /usr/lib/systemd/system/multipathd.service. 2026-03-09T17:45:07.135 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/sockets.target.wants/multipathd.socket → /usr/lib/systemd/system/multipathd.socket. 2026-03-09T17:45:07.135 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:07.669 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 135/146 2026-03-09T17:45:07.694 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 135/146 2026-03-09T17:45:07.694 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:45:07.695 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T17:45:07.695 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T17:45:07.695 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T17:45:07.695 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:07.706 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 136/146 2026-03-09T17:45:07.728 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 136/146 2026-03-09T17:45:07.728 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:45:07.728 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T17:45:07.728 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:07.889 INFO:teuthology.orchestra.run.vm05.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 137/146 2026-03-09T17:45:07.915 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 137/146 2026-03-09T17:45:07.915 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:45:07.915 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T17:45:07.915 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T17:45:07.915 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T17:45:07.915 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:08.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:08 vm01 ceph-mon[52793]: pgmap v158: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:08.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:08 vm04 ceph-mon[53158]: pgmap v158: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:08.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:08 vm05 ceph-mon[53831]: pgmap v158: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:10.514 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 138/146 2026-03-09T17:45:10.527 INFO:teuthology.orchestra.run.vm05.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 139/146 2026-03-09T17:45:10.534 INFO:teuthology.orchestra.run.vm05.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 140/146 2026-03-09T17:45:10.590 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 141/146 2026-03-09T17:45:10.600 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 142/146 2026-03-09T17:45:10.604 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 143/146 2026-03-09T17:45:10.604 INFO:teuthology.orchestra.run.vm05.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-09T17:45:10.619 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-09T17:45:10.619 INFO:teuthology.orchestra.run.vm05.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 145/146 2026-03-09T17:45:10.634 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 145/146 2026-03-09T17:45:10.652 INFO:teuthology.orchestra.run.vm05.stdout: Cleanup : kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-09T17:45:10.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:10 vm01 ceph-mon[52793]: pgmap v159: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:10.732 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:10 vm05 ceph-mon[53831]: pgmap v159: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:10.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:10 vm04 ceph-mon[53158]: pgmap v159: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:12.604 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-09T17:45:12.604 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/146 2026-03-09T17:45:12.604 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/146 2026-03-09T17:45:12.604 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/146 2026-03-09T17:45:12.604 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-09T17:45:12.604 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/146 2026-03-09T17:45:12.604 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/146 2026-03-09T17:45:12.604 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/146 2026-03-09T17:45:12.605 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : device-mapper-multipath-0.8.7-45.el9.x86_64 38/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : device-mapper-multipath-libs-0.8.7-45.el9.x86_64 39/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 40/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 41/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : isns-utils-libs-0.101-4.el9.x86_64 42/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 43/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 44/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 45/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 46/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 47/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 48/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 49/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 50/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-ply-3.11-14.el9.noarch 51/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 52/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 53/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 54/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 55/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : unzip-6.0-59.el9.x86_64 56/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : zip-3.0-35.el9.x86_64 57/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 58/146 2026-03-09T17:45:12.606 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 59/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 60/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 61/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 62/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 63/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 64/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 65/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 66/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 67/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 68/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : lua-5.4.4-4.el9.x86_64 69/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 70/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 71/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 72/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 73/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 74/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 75/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 76/146 2026-03-09T17:45:12.607 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 77/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 78/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 79/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 80/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 81/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 82/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 83/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 84/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 85/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 87/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 88/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 89/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 90/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 91/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 92/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 93/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 94/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 95/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 96/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 97/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 98/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 99/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 100/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 101/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 102/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 103/146 2026-03-09T17:45:12.608 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 104/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 105/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 106/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 107/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 108/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 109/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 110/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 111/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 112/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 113/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 114/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 115/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 116/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 117/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 118/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 119/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 120/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 121/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 122/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 123/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 124/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 125/146 2026-03-09T17:45:12.609 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 126/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 127/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 128/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 129/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 130/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 131/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 132/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 133/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 134/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 135/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 136/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 137/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 138/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : re2-1:20211101-20.el9.x86_64 139/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 140/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 141/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 142/146 2026-03-09T17:45:12.610 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 143/146 2026-03-09T17:45:12.611 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-09T17:45:12.611 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : kpartx-0.8.7-45.el9.x86_64 145/146 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout:Upgraded: 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: kpartx-0.8.7-45.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout:Installed: 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.723 INFO:teuthology.orchestra.run.vm05.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: device-mapper-multipath-0.8.7-45.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: device-mapper-multipath-libs-0.8.7-45.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3e853.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: isns-utils-libs-0.101-4.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.724 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyparsing-2.4.7-9.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.725 INFO:teuthology.orchestra.run.vm05.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.726 INFO:teuthology.orchestra.run.vm05.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:45:12.726 INFO:teuthology.orchestra.run.vm05.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T17:45:12.726 INFO:teuthology.orchestra.run.vm05.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T17:45:12.726 INFO:teuthology.orchestra.run.vm05.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T17:45:12.726 INFO:teuthology.orchestra.run.vm05.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T17:45:12.726 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T17:45:12.726 INFO:teuthology.orchestra.run.vm05.stdout: zip-3.0-35.el9.x86_64 2026-03-09T17:45:12.726 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:12.726 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:45:12.818 DEBUG:teuthology.parallel:result is None 2026-03-09T17:45:12.818 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T17:45:12.879 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:12 vm05 ceph-mon[53831]: pgmap v160: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:12.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:12 vm01 ceph-mon[52793]: pgmap v160: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:13.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:12 vm04 ceph-mon[53158]: pgmap v160: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:13.491 DEBUG:teuthology.orchestra.run.vm01:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T17:45:13.511 INFO:teuthology.orchestra.run.vm01.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T17:45:13.512 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T17:45:13.512 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T17:45:13.513 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T17:45:14.119 DEBUG:teuthology.orchestra.run.vm04:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T17:45:14.140 INFO:teuthology.orchestra.run.vm04.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T17:45:14.140 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T17:45:14.140 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T17:45:14.141 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T17:45:14.816 DEBUG:teuthology.orchestra.run.vm05:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T17:45:14.835 INFO:teuthology.orchestra.run.vm05.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T17:45:14.835 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T17:45:14.835 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T17:45:14.836 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-09T17:45:14.836 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:45:14.836 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T17:45:14.859 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:45:14 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:45:14.859 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:14 vm01 ceph-mon[52793]: pgmap v161: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:14.863 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:45:14.863 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T17:45:14.889 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:14 vm04 ceph-mon[53158]: pgmap v161: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:14.893 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-09T17:45:14.893 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T17:45:14.916 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:14 vm05 ceph-mon[53831]: pgmap v161: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:14.921 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-09T17:45:14.921 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:45:14.921 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T17:45:14.947 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T17:45:15.013 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:45:15.014 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T17:45:15.039 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T17:45:15.107 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-09T17:45:15.107 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T17:45:15.131 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T17:45:15.195 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-09T17:45:15.204 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:45:15.204 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T17:45:15.231 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T17:45:15.294 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:45:15.294 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T17:45:15.324 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T17:45:15.391 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-09T17:45:15.391 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T17:45:15.417 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T17:45:15.480 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-09T17:45:15.480 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-09T17:45:15.480 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T17:45:15.504 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T17:45:15.567 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:45:15.567 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T17:45:15.592 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T17:45:15.664 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-09T17:45:15.664 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T17:45:15.691 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T17:45:15.753 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:15 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:15.756 INFO:teuthology.run_tasks:Running task pexec... 2026-03-09T17:45:15.759 INFO:teuthology.task.pexec:Executing custom commands... 2026-03-09T17:45:15.759 DEBUG:teuthology.orchestra.run.vm01:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T17:45:15.759 DEBUG:teuthology.orchestra.run.vm04:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T17:45:15.760 DEBUG:teuthology.orchestra.run.vm05:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T17:45:15.761 DEBUG:teuthology.task.pexec:ubuntu@vm04.local< sudo dnf remove nvme-cli -y 2026-03-09T17:45:15.761 DEBUG:teuthology.task.pexec:ubuntu@vm04.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-09T17:45:15.761 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm04.local 2026-03-09T17:45:15.761 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T17:45:15.761 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-09T17:45:15.761 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo dnf remove nvme-cli -y 2026-03-09T17:45:15.761 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-09T17:45:15.762 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm01.local 2026-03-09T17:45:15.762 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T17:45:15.762 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-09T17:45:15.798 DEBUG:teuthology.task.pexec:ubuntu@vm05.local< sudo dnf remove nvme-cli -y 2026-03-09T17:45:15.798 DEBUG:teuthology.task.pexec:ubuntu@vm05.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-09T17:45:15.798 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm05.local 2026-03-09T17:45:15.798 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T17:45:15.798 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-09T17:45:15.810 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:15 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:15.956 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: nvme-cli 2026-03-09T17:45:15.956 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:45:15.960 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:45:15.960 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:45:15.960 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:45:15.973 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: nvme-cli 2026-03-09T17:45:15.973 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:45:15.976 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:45:15.977 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:45:15.977 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:45:16.000 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: nvme-cli 2026-03-09T17:45:16.000 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:45:16.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:15 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:16.003 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:45:16.004 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:45:16.004 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:45:16.026 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:45:15 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:45:16.343 INFO:teuthology.orchestra.run.vm01.stdout:Last metadata expiration check: 0:01:02 ago on Mon 09 Mar 2026 05:44:14 PM UTC. 2026-03-09T17:45:16.367 INFO:teuthology.orchestra.run.vm04.stdout:Last metadata expiration check: 0:00:50 ago on Mon 09 Mar 2026 05:44:26 PM UTC. 2026-03-09T17:45:16.375 INFO:teuthology.orchestra.run.vm05.stdout:Last metadata expiration check: 0:00:46 ago on Mon 09 Mar 2026 05:44:30 PM UTC. 2026-03-09T17:45:16.437 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout: Package Architecture Version Repository Size 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout:Installing: 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout:Installing dependencies: 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout:Install 5 Packages 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout:Total download size: 2.2 M 2026-03-09T17:45:16.438 INFO:teuthology.orchestra.run.vm01.stdout:Installed size: 11 M 2026-03-09T17:45:16.439 INFO:teuthology.orchestra.run.vm01.stdout:Downloading Packages: 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout: Package Architecture Version Repository Size 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout:Installing: 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout:Installing dependencies: 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout:Install 5 Packages 2026-03-09T17:45:16.462 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:45:16.463 INFO:teuthology.orchestra.run.vm04.stdout:Total download size: 2.2 M 2026-03-09T17:45:16.463 INFO:teuthology.orchestra.run.vm04.stdout:Installed size: 11 M 2026-03-09T17:45:16.463 INFO:teuthology.orchestra.run.vm04.stdout:Downloading Packages: 2026-03-09T17:45:16.470 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout: Package Architecture Version Repository Size 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout:Installing: 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout:Installing dependencies: 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout:Install 5 Packages 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout:Total download size: 2.2 M 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout:Installed size: 11 M 2026-03-09T17:45:16.471 INFO:teuthology.orchestra.run.vm05.stdout:Downloading Packages: 2026-03-09T17:45:16.681 INFO:teuthology.orchestra.run.vm04.stdout:(1/5): nvmetcli-0.8-3.el9.noarch.rpm 379 kB/s | 44 kB 00:00 2026-03-09T17:45:16.689 INFO:teuthology.orchestra.run.vm04.stdout:(2/5): python3-configshell-1.1.30-1.el9.noarch. 583 kB/s | 72 kB 00:00 2026-03-09T17:45:16.739 INFO:teuthology.orchestra.run.vm04.stdout:(3/5): python3-kmod-0.9-32.el9.x86_64.rpm 1.4 MB/s | 84 kB 00:00 2026-03-09T17:45:16.753 INFO:teuthology.orchestra.run.vm01.stdout:(1/5): nvmetcli-0.8-3.el9.noarch.rpm 207 kB/s | 44 kB 00:00 2026-03-09T17:45:16.778 INFO:teuthology.orchestra.run.vm01.stdout:(2/5): python3-configshell-1.1.30-1.el9.noarch. 304 kB/s | 72 kB 00:00 2026-03-09T17:45:16.797 INFO:teuthology.orchestra.run.vm04.stdout:(4/5): nvme-cli-2.16-1.el9.x86_64.rpm 5.0 MB/s | 1.2 MB 00:00 2026-03-09T17:45:16.805 INFO:teuthology.orchestra.run.vm04.stdout:(5/5): python3-urwid-2.1.2-4.el9.x86_64.rpm 7.1 MB/s | 837 kB 00:00 2026-03-09T17:45:16.805 INFO:teuthology.orchestra.run.vm04.stdout:-------------------------------------------------------------------------------- 2026-03-09T17:45:16.805 INFO:teuthology.orchestra.run.vm04.stdout:Total 6.3 MB/s | 2.2 MB 00:00 2026-03-09T17:45:16.848 INFO:teuthology.orchestra.run.vm05.stdout:(1/5): nvmetcli-0.8-3.el9.noarch.rpm 169 kB/s | 44 kB 00:00 2026-03-09T17:45:16.850 INFO:teuthology.orchestra.run.vm05.stdout:(2/5): python3-configshell-1.1.30-1.el9.noarch. 275 kB/s | 72 kB 00:00 2026-03-09T17:45:16.853 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T17:45:16.860 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T17:45:16.861 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T17:45:16.920 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T17:45:16.920 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T17:45:16.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:16 vm01 ceph-mon[52793]: pgmap v162: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:16.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:16 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:16.978 INFO:teuthology.orchestra.run.vm05.stdout:(3/5): python3-kmod-0.9-32.el9.x86_64.rpm 650 kB/s | 84 kB 00:00 2026-03-09T17:45:17.006 INFO:teuthology.orchestra.run.vm01.stdout:(3/5): nvme-cli-2.16-1.el9.x86_64.rpm 2.5 MB/s | 1.2 MB 00:00 2026-03-09T17:45:17.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:16 vm04 ceph-mon[53158]: pgmap v162: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:17.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:16 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:17.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:16 vm05 ceph-mon[53831]: pgmap v162: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:17.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:16 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:17.076 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T17:45:17.091 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/5 2026-03-09T17:45:17.093 INFO:teuthology.orchestra.run.vm01.stdout:(4/5): python3-kmod-0.9-32.el9.x86_64.rpm 248 kB/s | 84 kB 00:00 2026-03-09T17:45:17.099 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 2/5 2026-03-09T17:45:17.103 INFO:teuthology.orchestra.run.vm05.stdout:(4/5): nvme-cli-2.16-1.el9.x86_64.rpm 2.2 MB/s | 1.2 MB 00:00 2026-03-09T17:45:17.106 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 3/5 2026-03-09T17:45:17.107 INFO:teuthology.orchestra.run.vm04.stdout: Installing : nvmetcli-0.8-3.el9.noarch 4/5 2026-03-09T17:45:17.111 INFO:teuthology.orchestra.run.vm05.stdout:(5/5): python3-urwid-2.1.2-4.el9.x86_64.rpm 3.1 MB/s | 837 kB 00:00 2026-03-09T17:45:17.112 INFO:teuthology.orchestra.run.vm05.stdout:-------------------------------------------------------------------------------- 2026-03-09T17:45:17.112 INFO:teuthology.orchestra.run.vm05.stdout:Total 3.4 MB/s | 2.2 MB 00:00 2026-03-09T17:45:17.153 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-09T17:45:17.160 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-09T17:45:17.160 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-09T17:45:17.216 INFO:teuthology.orchestra.run.vm01.stdout:(5/5): python3-urwid-2.1.2-4.el9.x86_64.rpm 1.9 MB/s | 837 kB 00:00 2026-03-09T17:45:17.217 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-09T17:45:17.217 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-09T17:45:17.219 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-09T17:45:17.219 INFO:teuthology.orchestra.run.vm01.stdout:Total 2.8 MB/s | 2.2 MB 00:00 2026-03-09T17:45:17.256 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 4/5 2026-03-09T17:45:17.261 INFO:teuthology.orchestra.run.vm04.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 5/5 2026-03-09T17:45:17.266 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T17:45:17.274 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T17:45:17.274 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T17:45:17.334 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T17:45:17.334 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T17:45:17.376 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-09T17:45:17.391 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/5 2026-03-09T17:45:17.398 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 2/5 2026-03-09T17:45:17.404 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 3/5 2026-03-09T17:45:17.406 INFO:teuthology.orchestra.run.vm05.stdout: Installing : nvmetcli-0.8-3.el9.noarch 4/5 2026-03-09T17:45:17.491 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T17:45:17.506 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/5 2026-03-09T17:45:17.514 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 2/5 2026-03-09T17:45:17.521 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 3/5 2026-03-09T17:45:17.523 INFO:teuthology.orchestra.run.vm01.stdout: Installing : nvmetcli-0.8-3.el9.noarch 4/5 2026-03-09T17:45:17.553 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 4/5 2026-03-09T17:45:17.557 INFO:teuthology.orchestra.run.vm05.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 5/5 2026-03-09T17:45:17.636 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 5/5 2026-03-09T17:45:17.636 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T17:45:17.636 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:45:17.676 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 4/5 2026-03-09T17:45:17.680 INFO:teuthology.orchestra.run.vm01.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 5/5 2026-03-09T17:45:17.926 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 5/5 2026-03-09T17:45:17.927 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T17:45:17.927 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:18.032 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 5/5 2026-03-09T17:45:18.033 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T17:45:18.033 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:45:18.113 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/5 2026-03-09T17:45:18.114 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/5 2026-03-09T17:45:18.114 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/5 2026-03-09T17:45:18.114 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/5 2026-03-09T17:45:18.187 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 5/5 2026-03-09T17:45:18.187 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:45:18.187 INFO:teuthology.orchestra.run.vm04.stdout:Installed: 2026-03-09T17:45:18.187 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T17:45:18.187 INFO:teuthology.orchestra.run.vm04.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T17:45:18.187 INFO:teuthology.orchestra.run.vm04.stdout: python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T17:45:18.187 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:45:18.187 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:45:18.228 DEBUG:teuthology.parallel:result is None 2026-03-09T17:45:18.326 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/5 2026-03-09T17:45:18.326 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/5 2026-03-09T17:45:18.326 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/5 2026-03-09T17:45:18.326 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/5 2026-03-09T17:45:18.405 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 5/5 2026-03-09T17:45:18.405 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:18.405 INFO:teuthology.orchestra.run.vm05.stdout:Installed: 2026-03-09T17:45:18.405 INFO:teuthology.orchestra.run.vm05.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T17:45:18.405 INFO:teuthology.orchestra.run.vm05.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T17:45:18.405 INFO:teuthology.orchestra.run.vm05.stdout: python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T17:45:18.405 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:45:18.405 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:45:18.447 DEBUG:teuthology.parallel:result is None 2026-03-09T17:45:18.483 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/5 2026-03-09T17:45:18.483 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/5 2026-03-09T17:45:18.483 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/5 2026-03-09T17:45:18.483 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/5 2026-03-09T17:45:18.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 5/5 2026-03-09T17:45:18.550 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:45:18.550 INFO:teuthology.orchestra.run.vm01.stdout:Installed: 2026-03-09T17:45:18.550 INFO:teuthology.orchestra.run.vm01.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T17:45:18.550 INFO:teuthology.orchestra.run.vm01.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T17:45:18.550 INFO:teuthology.orchestra.run.vm01.stdout: python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T17:45:18.550 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:45:18.550 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:45:18.593 DEBUG:teuthology.parallel:result is None 2026-03-09T17:45:18.593 INFO:teuthology.run_tasks:Running task ceph_iscsi_client... 2026-03-09T17:45:18.596 INFO:tasks.ceph_iscsi_client:Setting up ceph-iscsi client... 2026-03-09T17:45:18.596 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:45:18.596 DEBUG:teuthology.orchestra.run.vm04:> sudo mkdir -p /etc/iscsi 2026-03-09T17:45:18.596 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/iscsi/initiatorname.iscsi 2026-03-09T17:45:18.643 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl restart iscsid 2026-03-09T17:45:18.652 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:18 vm04 ceph-mon[53158]: pgmap v163: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:18.715 DEBUG:teuthology.orchestra.run.vm04:> sudo modprobe dm_multipath 2026-03-09T17:45:18.747 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-09T17:45:18.747 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/multipath.conf 2026-03-09T17:45:18.817 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl start multipathd 2026-03-09T17:45:18.938 INFO:teuthology.run_tasks:Running task cram... 2026-03-09T17:45:18.941 INFO:tasks.cram:Pulling tests from https://github.com/kshtsk/ceph.git ref 569c3e99c9b32a51b4eaf08731c728f4513ed589 2026-03-09T17:45:18.942 DEBUG:teuthology.orchestra.run.vm01:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.0 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-09T17:45:18.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:18 vm01 ceph-mon[52793]: pgmap v163: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:19.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:18 vm05 ceph-mon[53831]: pgmap v163: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:20.444 INFO:teuthology.orchestra.run.vm01.stdout:Collecting cram==0.6 2026-03-09T17:45:20.473 INFO:teuthology.orchestra.run.vm01.stdout: Downloading cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-09T17:45:20.487 INFO:teuthology.orchestra.run.vm01.stdout:Installing collected packages: cram 2026-03-09T17:45:20.494 INFO:teuthology.orchestra.run.vm01.stdout:Successfully installed cram-0.6 2026-03-09T17:45:20.578 INFO:teuthology.orchestra.run.vm01.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-03-09T17:45:20.578 INFO:teuthology.orchestra.run.vm01.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/virtualenv/bin/python3 -m pip install --upgrade pip' command. 2026-03-09T17:45:20.612 DEBUG:teuthology.orchestra.run.vm01:> rm -rf /home/ubuntu/cephtest/clone.client.0 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.0 && cd /home/ubuntu/cephtest/clone.client.0 && git checkout 569c3e99c9b32a51b4eaf08731c728f4513ed589 2026-03-09T17:45:20.672 INFO:teuthology.orchestra.run.vm01.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.0'... 2026-03-09T17:45:20.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:20 vm01 ceph-mon[52793]: pgmap v164: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:21.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:20 vm04 ceph-mon[53158]: pgmap v164: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:21.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:20 vm05 ceph-mon[53831]: pgmap v164: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:22.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:22 vm01 ceph-mon[52793]: pgmap v165: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:23.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:22 vm04 ceph-mon[53158]: pgmap v165: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:23.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:22 vm05 ceph-mon[53831]: pgmap v165: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:24 vm01 ceph-mon[52793]: pgmap v166: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:24.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:45:24 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:45:25.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:24 vm04 ceph-mon[53158]: pgmap v166: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:25.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:24 vm05 ceph-mon[53831]: pgmap v166: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:25.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:25 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:26.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:25 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:26.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:25 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:26.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:45:25 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:45:27.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:26 vm04 ceph-mon[53158]: pgmap v167: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:27.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:26 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:27.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:26 vm05 ceph-mon[53831]: pgmap v167: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:27.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:26 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:27.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:26 vm01 ceph-mon[52793]: pgmap v167: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:27.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:26 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:29.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:28 vm04 ceph-mon[53158]: pgmap v168: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:29.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:28 vm05 ceph-mon[53831]: pgmap v168: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:29.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:28 vm01 ceph-mon[52793]: pgmap v168: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:31.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:30 vm04 ceph-mon[53158]: pgmap v169: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:31.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:30 vm05 ceph-mon[53831]: pgmap v169: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:31.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:30 vm01 ceph-mon[52793]: pgmap v169: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:33.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:32 vm04 ceph-mon[53158]: pgmap v170: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:33.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:32 vm05 ceph-mon[53831]: pgmap v170: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:33.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:32 vm01 ceph-mon[52793]: pgmap v170: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:34.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:45:34 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:45:34.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:34 vm01 ceph-mon[52793]: pgmap v171: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:35.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:34 vm05 ceph-mon[53831]: pgmap v171: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:34 vm04 ceph-mon[53158]: pgmap v171: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:36.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:35 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:35 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:36.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:35 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:36.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:45:35 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:45:37.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:36 vm01 ceph-mon[52793]: pgmap v172: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:37.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:36 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:36 vm04 ceph-mon[53158]: pgmap v172: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:36 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:37.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:36 vm05 ceph-mon[53831]: pgmap v172: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:37.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:36 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:38.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:45:38.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:45:38.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:45:38.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:45:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:37 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:45:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:37 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:45:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:37 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:45:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:37 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:45:38.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:37 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:45:38.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:37 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:45:38.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:37 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:45:38.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:37 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:45:39.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:38 vm01 ceph-mon[52793]: pgmap v173: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:38 vm04 ceph-mon[53158]: pgmap v173: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:38 vm05 ceph-mon[53831]: pgmap v173: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:40.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:40 vm04 ceph-mon[53158]: pgmap v174: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:40.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:40 vm05 ceph-mon[53831]: pgmap v174: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:40.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:40 vm01 ceph-mon[52793]: pgmap v174: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:42.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:42 vm01 ceph-mon[52793]: pgmap v175: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:42.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:42 vm04 ceph-mon[53158]: pgmap v175: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:42.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:42 vm05 ceph-mon[53831]: pgmap v175: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:44.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:44 vm04 ceph-mon[53158]: pgmap v176: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:44.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:44 vm05 ceph-mon[53831]: pgmap v176: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:44.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:44 vm01 ceph-mon[52793]: pgmap v176: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:44.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:45:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:45:45.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:45 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:45.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:45 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:45.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:45 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:46.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:45:45 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:45:46.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:46 vm04 ceph-mon[53158]: pgmap v177: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:46.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:46 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:46.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:46 vm05 ceph-mon[53831]: pgmap v177: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:46.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:46 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:46.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:46 vm01 ceph-mon[52793]: pgmap v177: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:46.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:46 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:48.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:48 vm04 ceph-mon[53158]: pgmap v178: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:48.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:48 vm05 ceph-mon[53831]: pgmap v178: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:48.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:48 vm01 ceph-mon[52793]: pgmap v178: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:50.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:50 vm04 ceph-mon[53158]: pgmap v179: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:50.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:50 vm05 ceph-mon[53831]: pgmap v179: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:50.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:50 vm01 ceph-mon[52793]: pgmap v179: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:52.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:52 vm04 ceph-mon[53158]: pgmap v180: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:52.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:52 vm05 ceph-mon[53831]: pgmap v180: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:52.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:52 vm01 ceph-mon[52793]: pgmap v180: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:54.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:54 vm04 ceph-mon[53158]: pgmap v181: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:54.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:54 vm05 ceph-mon[53831]: pgmap v181: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:54.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:54 vm01 ceph-mon[52793]: pgmap v181: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:54.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:45:54 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:45:55.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:55 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:55.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:55 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:55.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:55 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:56.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:45:55 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:45:56.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:56 vm04 ceph-mon[53158]: pgmap v182: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:56.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:56 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:56.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:56 vm05 ceph-mon[53831]: pgmap v182: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:56.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:56 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:56.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:56 vm01 ceph-mon[52793]: pgmap v182: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:45:56.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:56 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:45:58.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:45:58 vm04 ceph-mon[53158]: pgmap v183: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:58.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:45:58 vm05 ceph-mon[53831]: pgmap v183: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:45:58.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:45:58 vm01 ceph-mon[52793]: pgmap v183: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:00.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:00 vm04 ceph-mon[53158]: pgmap v184: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:00.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:00 vm05 ceph-mon[53831]: pgmap v184: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:00.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:00 vm01 ceph-mon[52793]: pgmap v184: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:02.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:02 vm04 ceph-mon[53158]: pgmap v185: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:02.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:02 vm05 ceph-mon[53831]: pgmap v185: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:02.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:02 vm01 ceph-mon[52793]: pgmap v185: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:04.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:04 vm01 ceph-mon[52793]: pgmap v186: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:04.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:46:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:46:05.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:04 vm04 ceph-mon[53158]: pgmap v186: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:05.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:04 vm05 ceph-mon[53831]: pgmap v186: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:05.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:05 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:06.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:05 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:06.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:05 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:06.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:46:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr:Note: switching to '569c3e99c9b32a51b4eaf08731c728f4513ed589'. 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr:state without impacting any branches by switching back to a branch. 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr: git switch -c 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr:Or undo this operation with: 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr: git switch - 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:46:06.228 INFO:teuthology.orchestra.run.vm01.stderr:HEAD is now at 569c3e99c9b qa/rgw: bucket notifications use pynose 2026-03-09T17:46:06.234 DEBUG:teuthology.orchestra.run.vm01:> cp -- /home/ubuntu/cephtest/clone.client.0/src/test/cli-integration/rbd/gwcli_create.t /home/ubuntu/cephtest/archive/cram.client.0 2026-03-09T17:46:06.291 DEBUG:teuthology.orchestra.run.vm04:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.1 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-09T17:46:06.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:06 vm01 ceph-mon[52793]: pgmap v187: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:06.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:06 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:07.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:06 vm04 ceph-mon[53158]: pgmap v187: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:07.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:06 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:07.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:06 vm05 ceph-mon[53831]: pgmap v187: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:07.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:06 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:07.698 INFO:teuthology.orchestra.run.vm04.stdout:Collecting cram==0.6 2026-03-09T17:46:07.728 INFO:teuthology.orchestra.run.vm04.stdout: Downloading cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-09T17:46:07.739 INFO:teuthology.orchestra.run.vm04.stdout:Installing collected packages: cram 2026-03-09T17:46:07.746 INFO:teuthology.orchestra.run.vm04.stdout:Successfully installed cram-0.6 2026-03-09T17:46:07.831 INFO:teuthology.orchestra.run.vm04.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-03-09T17:46:07.831 INFO:teuthology.orchestra.run.vm04.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/virtualenv/bin/python3 -m pip install --upgrade pip' command. 2026-03-09T17:46:07.865 DEBUG:teuthology.orchestra.run.vm04:> rm -rf /home/ubuntu/cephtest/clone.client.1 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.1 && cd /home/ubuntu/cephtest/clone.client.1 && git checkout 569c3e99c9b32a51b4eaf08731c728f4513ed589 2026-03-09T17:46:07.881 INFO:teuthology.orchestra.run.vm04.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.1'... 2026-03-09T17:46:08.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:08 vm01 ceph-mon[52793]: pgmap v188: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:09.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:08 vm04 ceph-mon[53158]: pgmap v188: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:09.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:08 vm05 ceph-mon[53831]: pgmap v188: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:10.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:10 vm01 ceph-mon[52793]: pgmap v189: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:11.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:10 vm04 ceph-mon[53158]: pgmap v189: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:11.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:10 vm05 ceph-mon[53831]: pgmap v189: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:12.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:12 vm01 ceph-mon[52793]: pgmap v190: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:13.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:12 vm04 ceph-mon[53158]: pgmap v190: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:13.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:12 vm05 ceph-mon[53831]: pgmap v190: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:14.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:46:14 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:46:14.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:14 vm01 ceph-mon[52793]: pgmap v191: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:15.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:14 vm04 ceph-mon[53158]: pgmap v191: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:15.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:14 vm05 ceph-mon[53831]: pgmap v191: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:15.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:15 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:16.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:15 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:16.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:15 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:16.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:46:15 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:46:17.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:16 vm04 ceph-mon[53158]: pgmap v192: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:17.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:16 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:17.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:16 vm05 ceph-mon[53831]: pgmap v192: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:17.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:16 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:17.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:16 vm01 ceph-mon[52793]: pgmap v192: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:17.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:16 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:19.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:18 vm04 ceph-mon[53158]: pgmap v193: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:19.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:18 vm05 ceph-mon[53831]: pgmap v193: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:19.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:18 vm01 ceph-mon[52793]: pgmap v193: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:21.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:20 vm05 ceph-mon[53831]: pgmap v194: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:21.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:20 vm01 ceph-mon[52793]: pgmap v194: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:21.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:20 vm04 ceph-mon[53158]: pgmap v194: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:23.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:22 vm05 ceph-mon[53831]: pgmap v195: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:23.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:22 vm01 ceph-mon[52793]: pgmap v195: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:23.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:22 vm04 ceph-mon[53158]: pgmap v195: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:25.009 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:46:24 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:46:25.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:25 vm05 ceph-mon[53831]: pgmap v196: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:25.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:25 vm01 ceph-mon[52793]: pgmap v196: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:25.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:25 vm04 ceph-mon[53158]: pgmap v196: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:26.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:26 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:26.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:26 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:26.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:46:25 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:46:26.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:26 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:27.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:27 vm04 ceph-mon[53158]: pgmap v197: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:27.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:27 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:27.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:27 vm05 ceph-mon[53831]: pgmap v197: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:27.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:27 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:27.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:27 vm01 ceph-mon[52793]: pgmap v197: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:27.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:27 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:29.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:29 vm04 ceph-mon[53158]: pgmap v198: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:29.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:29 vm05 ceph-mon[53831]: pgmap v198: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:29.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:29 vm01 ceph-mon[52793]: pgmap v198: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:31.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:31 vm05 ceph-mon[53831]: pgmap v199: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:31.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:31 vm01 ceph-mon[52793]: pgmap v199: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:31.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:31 vm04 ceph-mon[53158]: pgmap v199: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:33.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:33 vm05 ceph-mon[53831]: pgmap v200: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:33.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:33 vm01 ceph-mon[52793]: pgmap v200: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:33.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:33 vm04 ceph-mon[53158]: pgmap v200: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:35.057 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:46:34 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:46:35.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:35 vm05 ceph-mon[53831]: pgmap v201: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:35.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:35 vm01 ceph-mon[52793]: pgmap v201: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:35.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:35 vm04 ceph-mon[53158]: pgmap v201: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:36.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:46:35 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:46:36.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:36 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:36.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:36 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:36.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:36 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:37 vm01 ceph-mon[52793]: pgmap v202: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:37 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:37.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:46:38.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:37 vm04 ceph-mon[53158]: pgmap v202: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:38.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:37 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:38.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:37 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:46:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:37 vm05 ceph-mon[53831]: pgmap v202: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:37 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:37 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:46:38.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:38 vm01 ceph-mon[52793]: pgmap v203: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:38.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:46:38.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:46:38.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:46:39.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:38 vm04 ceph-mon[53158]: pgmap v203: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:39.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:46:39.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:46:39.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:46:39.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:38 vm05 ceph-mon[53831]: pgmap v203: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:39.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:46:39.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:46:39.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:46:40.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:40 vm01 ceph-mon[52793]: pgmap v204: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:41.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:40 vm04 ceph-mon[53158]: pgmap v204: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:41.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:40 vm05 ceph-mon[53831]: pgmap v204: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:42.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:42 vm01 ceph-mon[52793]: pgmap v205: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:43.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:42 vm04 ceph-mon[53158]: pgmap v205: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:43.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:42 vm05 ceph-mon[53831]: pgmap v205: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:44.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:44 vm01 ceph-mon[52793]: pgmap v206: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:44.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:46:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:46:45.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:44 vm04 ceph-mon[53158]: pgmap v206: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:45.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:44 vm05 ceph-mon[53831]: pgmap v206: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:46.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:45 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:46.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:45 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:46.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:46:45 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:46:46.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:45 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:47.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:46 vm04 ceph-mon[53158]: pgmap v207: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:47.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:46 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:47.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:46 vm05 ceph-mon[53831]: pgmap v207: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:47.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:46 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:47.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:46 vm01 ceph-mon[52793]: pgmap v207: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:47.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:46 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr:Note: switching to '569c3e99c9b32a51b4eaf08731c728f4513ed589'. 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr:state without impacting any branches by switching back to a branch. 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr: git switch -c 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr:Or undo this operation with: 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr: git switch - 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:46:48.882 INFO:teuthology.orchestra.run.vm04.stderr:HEAD is now at 569c3e99c9b qa/rgw: bucket notifications use pynose 2026-03-09T17:46:48.888 DEBUG:teuthology.orchestra.run.vm04:> cp -- /home/ubuntu/cephtest/clone.client.1/src/test/cli-integration/rbd/iscsi_client.t /home/ubuntu/cephtest/archive/cram.client.1 2026-03-09T17:46:48.944 DEBUG:teuthology.orchestra.run.vm05:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.2 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-09T17:46:49.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:48 vm04 ceph-mon[53158]: pgmap v208: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:49.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:48 vm05 ceph-mon[53831]: pgmap v208: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:49.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:48 vm01 ceph-mon[52793]: pgmap v208: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:50.319 INFO:teuthology.orchestra.run.vm05.stdout:Collecting cram==0.6 2026-03-09T17:46:50.348 INFO:teuthology.orchestra.run.vm05.stdout: Downloading cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-09T17:46:50.360 INFO:teuthology.orchestra.run.vm05.stdout:Installing collected packages: cram 2026-03-09T17:46:50.366 INFO:teuthology.orchestra.run.vm05.stdout:Successfully installed cram-0.6 2026-03-09T17:46:50.446 INFO:teuthology.orchestra.run.vm05.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-03-09T17:46:50.446 INFO:teuthology.orchestra.run.vm05.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/virtualenv/bin/python3 -m pip install --upgrade pip' command. 2026-03-09T17:46:50.476 DEBUG:teuthology.orchestra.run.vm05:> rm -rf /home/ubuntu/cephtest/clone.client.2 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.2 && cd /home/ubuntu/cephtest/clone.client.2 && git checkout 569c3e99c9b32a51b4eaf08731c728f4513ed589 2026-03-09T17:46:50.491 INFO:teuthology.orchestra.run.vm05.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.2'... 2026-03-09T17:46:51.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:50 vm04 ceph-mon[53158]: pgmap v209: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:51.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:50 vm05 ceph-mon[53831]: pgmap v209: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:51.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:50 vm01 ceph-mon[52793]: pgmap v209: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:53.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:52 vm04 ceph-mon[53158]: pgmap v210: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:53.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:52 vm05 ceph-mon[53831]: pgmap v210: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:53.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:52 vm01 ceph-mon[52793]: pgmap v210: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:55.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:54 vm05 ceph-mon[53831]: pgmap v211: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:55.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:46:54 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:46:55.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:54 vm01 ceph-mon[52793]: pgmap v211: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:55.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:54 vm04 ceph-mon[53158]: pgmap v211: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:56.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:55 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:56.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:46:55 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:46:56.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:55 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:56.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:55 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:57.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:56 vm05 ceph-mon[53831]: pgmap v212: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:57.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:56 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:57.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:56 vm01 ceph-mon[52793]: pgmap v212: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:57.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:56 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:57.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:56 vm04 ceph-mon[53158]: pgmap v212: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:46:57.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:56 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:46:59.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:46:58 vm01 ceph-mon[52793]: pgmap v213: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:46:58 vm04 ceph-mon[53158]: pgmap v213: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:46:59.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:46:58 vm05 ceph-mon[53831]: pgmap v213: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:01.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:00 vm01 ceph-mon[52793]: pgmap v214: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:01.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:00 vm04 ceph-mon[53158]: pgmap v214: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:01.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:00 vm05 ceph-mon[53831]: pgmap v214: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:03.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:02 vm01 ceph-mon[52793]: pgmap v215: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:03.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:02 vm04 ceph-mon[53158]: pgmap v215: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:03.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:02 vm05 ceph-mon[53831]: pgmap v215: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:05.096 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:47:05.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:05 vm01 ceph-mon[52793]: pgmap v216: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:05 vm04 ceph-mon[53158]: pgmap v216: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:05 vm05 ceph-mon[53831]: pgmap v216: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:06.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:47:06.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:06 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:06.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:06 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:06.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:06 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:07.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:07 vm01 ceph-mon[52793]: pgmap v217: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:07.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:07 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:07.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:07 vm04 ceph-mon[53158]: pgmap v217: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:07.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:07 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:07.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:07 vm05 ceph-mon[53831]: pgmap v217: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:07.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:07 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:09.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:09 vm01 ceph-mon[52793]: pgmap v218: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:09.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:09 vm04 ceph-mon[53158]: pgmap v218: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:09.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:09 vm05 ceph-mon[53831]: pgmap v218: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:11.136 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:11 vm04 ceph-mon[53158]: pgmap v219: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:11.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:11 vm01 ceph-mon[52793]: pgmap v219: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:11.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:11 vm05 ceph-mon[53831]: pgmap v219: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:13.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:13 vm01 ceph-mon[52793]: pgmap v220: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:13.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:13 vm04 ceph-mon[53158]: pgmap v220: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:13.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:13 vm05 ceph-mon[53831]: pgmap v220: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:15.148 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:14 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:47:15.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:15 vm01 ceph-mon[52793]: pgmap v221: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:15.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:15 vm04 ceph-mon[53158]: pgmap v221: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:15.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:15 vm05 ceph-mon[53831]: pgmap v221: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:16.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:16 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:16.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:15 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:47:16.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:16 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:16.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:16 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:17.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:17 vm01 ceph-mon[52793]: pgmap v222: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:17.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:17 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:17.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:17 vm04 ceph-mon[53158]: pgmap v222: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:17.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:17 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:17.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:17 vm05 ceph-mon[53831]: pgmap v222: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:17.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:17 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:18.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:18 vm01 ceph-mon[52793]: pgmap v223: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:18.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:18 vm04 ceph-mon[53158]: pgmap v223: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:18.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:18 vm05 ceph-mon[53831]: pgmap v223: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:20.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:20 vm04 ceph-mon[53158]: pgmap v224: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:20.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:20 vm05 ceph-mon[53831]: pgmap v224: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:20.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:20 vm01 ceph-mon[52793]: pgmap v224: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:22.904 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:22 vm05 ceph-mon[53831]: pgmap v225: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:22.933 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:22 vm01 ceph-mon[52793]: pgmap v225: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:23.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:22 vm04 ceph-mon[53158]: pgmap v225: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:24.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:24 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:47:24.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:24 vm01 ceph-mon[52793]: pgmap v226: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:25.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:24 vm04 ceph-mon[53158]: pgmap v226: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:25.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:24 vm05 ceph-mon[53831]: pgmap v226: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:25.966 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:25 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:25.967 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:25 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:47:25.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:25 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:26.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:25 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:26.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:26 vm01 ceph-mon[52793]: pgmap v227: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:26.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:26 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:27.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:26 vm04 ceph-mon[53158]: pgmap v227: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:27.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:26 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:27.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:26 vm05 ceph-mon[53831]: pgmap v227: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:27.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:26 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:28.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:28 vm01 ceph-mon[52793]: pgmap v228: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:29.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:28 vm04 ceph-mon[53158]: pgmap v228: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:29.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:28 vm05 ceph-mon[53831]: pgmap v228: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:30.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:30 vm01 ceph-mon[52793]: pgmap v229: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:31.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:30 vm04 ceph-mon[53158]: pgmap v229: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:31.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:30 vm05 ceph-mon[53831]: pgmap v229: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:32.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:32 vm01 ceph-mon[52793]: pgmap v230: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:33.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:32 vm04 ceph-mon[53158]: pgmap v230: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:33.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:32 vm05 ceph-mon[53831]: pgmap v230: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:35.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:34 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:47:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:34 vm01 ceph-mon[52793]: pgmap v231: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:34 vm04 ceph-mon[53158]: pgmap v231: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:35.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:35 vm05 ceph-mon[53831]: pgmap v231: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:36.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:35 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:35 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:36.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:35 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:36.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:35 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:47:37.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:36 vm01 ceph-mon[52793]: pgmap v232: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:37.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:36 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:36 vm04 ceph-mon[53158]: pgmap v232: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:36 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:37.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:36 vm05 ceph-mon[53831]: pgmap v232: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:37.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:36 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:37 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:47:38.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:37 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:47:38.450 INFO:teuthology.orchestra.run.vm05.stderr:Note: switching to '569c3e99c9b32a51b4eaf08731c728f4513ed589'. 2026-03-09T17:47:38.450 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:47:38.450 INFO:teuthology.orchestra.run.vm05.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-09T17:47:38.450 INFO:teuthology.orchestra.run.vm05.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-09T17:47:38.450 INFO:teuthology.orchestra.run.vm05.stderr:state without impacting any branches by switching back to a branch. 2026-03-09T17:47:38.451 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:47:38.451 INFO:teuthology.orchestra.run.vm05.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-09T17:47:38.451 INFO:teuthology.orchestra.run.vm05.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-09T17:47:38.451 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:47:38.451 INFO:teuthology.orchestra.run.vm05.stderr: git switch -c 2026-03-09T17:47:38.451 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:47:38.451 INFO:teuthology.orchestra.run.vm05.stderr:Or undo this operation with: 2026-03-09T17:47:38.451 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:47:38.451 INFO:teuthology.orchestra.run.vm05.stderr: git switch - 2026-03-09T17:47:38.451 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:47:38.451 INFO:teuthology.orchestra.run.vm05.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-09T17:47:38.451 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:47:38.451 INFO:teuthology.orchestra.run.vm05.stderr:HEAD is now at 569c3e99c9b qa/rgw: bucket notifications use pynose 2026-03-09T17:47:38.457 DEBUG:teuthology.orchestra.run.vm05:> cp -- /home/ubuntu/cephtest/clone.client.2/src/test/cli-integration/rbd/gwcli_delete.t /home/ubuntu/cephtest/archive/cram.client.2 2026-03-09T17:47:38.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:37 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:47:38.512 INFO:tasks.cram:Running tests for client.0... 2026-03-09T17:47:38.513 DEBUG:teuthology.orchestra.run.vm01:> CEPH_REF=master CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.0/*.t 2026-03-09T17:47:39.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:38 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:38] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:39.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:38 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:38] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:39.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:38 vm01 ceph-mon[52793]: pgmap v233: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:39.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:47:39.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:47:39.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:38 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:47:39.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:38 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3275403222' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:39.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:38 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2600756086' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:39.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:38 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2958051963' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:38 vm04 ceph-mon[53158]: pgmap v233: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:47:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:47:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:38 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:47:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:38 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3275403222' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:38 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2600756086' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:38 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2958051963' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:38 vm05 ceph-mon[53831]: pgmap v233: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:47:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:47:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:47:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:38 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:47:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:38 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3275403222' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:38 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2600756086' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:39.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:38 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2958051963' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:39] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:39] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:39] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:39] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (LUN.allocate) created datapool/block0 successfully 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block0' to LIO backstore user:rbd 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: /usr/lib/python3.9/site-packages/rtslib_fb/root.py:180: UserWarning: Cannot set dbroot to /var/target. Target devices have already been registered. 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: warn("Cannot set dbroot to {}. Target devices have already been registered." 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (LUN.add_dev_to_lio) Successfully added datapool/block0 to LIO 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (LUN.allocate) added 'datapool/block0' to LIO and config object 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:39] "PUT /api/_disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:39] "PUT /api/_disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _disk update on localhost, successful 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:39] "PUT /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:39] "PUT /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:39] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:39] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:39] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:39] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:39] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:39] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:39] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:39.658 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:39] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:40.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:40 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:40] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:40.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:40 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:40] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:40.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:40 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:40] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:40.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:40 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:40] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:40.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:40 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:40] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:40.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:40 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:40] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:40.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3899374284' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:40.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3771931189' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:40.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1496417834' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:40.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/626404148' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:40.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3229755844' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:40.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1364058674' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:40.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3256324872' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:40.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3794828077' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:47:40.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3794828077' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:47:40.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3794828077' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:47:40.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:39 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/16777295' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:39 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3899374284' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:39 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3771931189' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:39 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1496417834' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:39 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/626404148' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:39 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3229755844' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:39 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1364058674' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:39 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3256324872' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:39 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3794828077' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:47:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:39 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3794828077' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:47:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:39 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3794828077' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:47:40.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:39 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/16777295' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:40.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:39 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3899374284' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:40.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:39 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3771931189' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:40.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:39 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1496417834' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:40.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:39 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/626404148' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:40.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:39 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3229755844' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:40.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:39 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1364058674' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:40.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:39 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3256324872' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:40.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:39 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3794828077' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:47:40.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:39 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3794828077' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:47:40.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:39 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3794828077' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:47:40.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:39 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/16777295' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:40.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:40 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug Change detected - internal 0 / xattr 1 refreshing 2026-03-09T17:47:40.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:40 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:40] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:40.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:40 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:40] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:40.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:40 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:40] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:40.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:40 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:40] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:40.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:40 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:40] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:40.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:40 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:40] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:41.268 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:41] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:41.268 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:41] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:41.268 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:41] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:41.268 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:41] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:41.268 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:41] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:41.268 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:41] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:41.268 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-mon[52793]: pgmap v234: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:41.268 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2949045873' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:41.268 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2816048633' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:41.268 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/912370428' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:41.268 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3554115957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:41.268 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3077111945' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:41.268 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/235281594' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:41.268 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3749471588' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:41.268 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2828449261' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:41.268 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/893879720' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:41.268 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/522880171' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:41 vm04 ceph-mon[53158]: pgmap v234: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:41 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2949045873' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:41 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2816048633' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:41 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/912370428' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:41 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3554115957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:41 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3077111945' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:41 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/235281594' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:41 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3749471588' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:41 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2828449261' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:41 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/893879720' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:41 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/522880171' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:41 vm05 ceph-mon[53831]: pgmap v234: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:47:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:41 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2949045873' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:41 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2816048633' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:41 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/912370428' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:41 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3554115957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:41 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3077111945' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:41 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/235281594' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:41.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:41 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3749471588' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:41.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:41 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2828449261' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:41.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:41 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/893879720' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:41.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:41 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/522880171' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:41.908 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:41] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:41.908 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:41] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:41.908 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:41] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:41.908 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:41] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:41.908 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Gateway.create_target) created an iscsi target with iqn of 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-09T17:47:41.908 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:41] "PUT /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:47:41.908 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:41] "PUT /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:47:41.908 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:41] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:41.908 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:41] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:41.908 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:41] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:41.908 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:41 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:41] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:42.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3795431631' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:42.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2807557837' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:42.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2297465675' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:42.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3408754861' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:42.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3013564170' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1020735935' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1832461877' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1712590887' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1142750874' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:42.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3816037664' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3795431631' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2807557837' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2297465675' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3408754861' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3013564170' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1020735935' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1832461877' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1712590887' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1142750874' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:42.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:42 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3816037664' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3795431631' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2807557837' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2297465675' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3408754861' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3013564170' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:42.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1020735935' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:42.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1832461877' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:42.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1712590887' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:42.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1142750874' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:42.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:42 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3816037664' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:42.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:42 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug Change detected - internal 1 / xattr 2 refreshing 2026-03-09T17:47:42.707 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:42] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:42.707 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:42] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:42.707 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:42] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:42.707 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:42] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:42.707 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:42] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:42.707 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:42] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:42.707 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:42] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:42.707 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:42] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:42.977 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:42] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:42.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:42] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:42.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:42] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:42.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:42] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:42.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:42] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:42.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:42] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:42.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:42] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:42.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:42 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:42] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:43.276 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:43] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:43.276 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:43] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:43.276 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:43] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:43.276 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:43] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:43.277 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-mon[52793]: pgmap v235: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 6.5 KiB/s rd, 426 B/s wr, 9 op/s 2026-03-09T17:47:43.277 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2673548971' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:43.277 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3602941953' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:43.277 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3736528126' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:43.277 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3618665183' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:43.277 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1139789529' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:43.277 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3976742430' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:43.277 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2527352645' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:43.277 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1391394117' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:43.277 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/375103947' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:43.277 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/820382433' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:43 vm04 ceph-mon[53158]: pgmap v235: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 6.5 KiB/s rd, 426 B/s wr, 9 op/s 2026-03-09T17:47:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:43 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2673548971' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:43 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3602941953' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:43 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3736528126' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:43 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3618665183' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:43 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1139789529' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:43 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3976742430' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:43 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2527352645' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:43 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1391394117' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:43 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/375103947' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:43 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/820382433' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:43 vm05 ceph-mon[53831]: pgmap v235: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 6.5 KiB/s rd, 426 B/s wr, 9 op/s 2026-03-09T17:47:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:43 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2673548971' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:43 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3602941953' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:43 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3736528126' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:43 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3618665183' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:43 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1139789529' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:43 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3976742430' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:43 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2527352645' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:43 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1391394117' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:43 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/375103947' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:43 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/820382433' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:43.681 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:43] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:43.681 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:43] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:43.681 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:43] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:43.681 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:43] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:43.977 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:43] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:43.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:43] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:43.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:43] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:43.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:43] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:43.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:43] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:43.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:43] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:43.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:43] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:43.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:43 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:43] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:44.285 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:44] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:44.285 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:44] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:44.285 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:44] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:44.285 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:44] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:44.285 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:44] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:44.285 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:44] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:44.285 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:44] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:44.285 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:44] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:44.285 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1970058346' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:44.285 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2697687747' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:44.285 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3792437789' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:44.285 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/286157843' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:44.285 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2001384153' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:44.285 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2835828808' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:44.285 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2198049602' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:44.285 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/638391538' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:44.285 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1371536726' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:44.285 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1027838064' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:44.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1970058346' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:44.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2697687747' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:44.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3792437789' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:44.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/286157843' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:44.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2001384153' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:44.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2835828808' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:44.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2198049602' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:44.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/638391538' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:44.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1371536726' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:44.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:44 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1027838064' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1970058346' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2697687747' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3792437789' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/286157843' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2001384153' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2835828808' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2198049602' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/638391538' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1371536726' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:44.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:44 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1027838064' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:44.977 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:44] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:44.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:44] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:44.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:44] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:44.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:44] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:44.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:44] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:44.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:44] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:44.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:44] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:44.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:44] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:44.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:47:45.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:45] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:45.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:45] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:45.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.278 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.278 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.278 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.278 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.278 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.278 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/sysinfo/hostname HTTP/1.1" 200 - 2026-03-09T17:47:45.278 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/sysinfo/hostname HTTP/1.1" 200 - 2026-03-09T17:47:45.278 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug gateway validation needed for vm01.local 2026-03-09T17:47:45.278 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-09T17:47:45.278 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-09T17:47:45.278 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-09T17:47:45.278 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-09T17:47:45.278 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-mon[52793]: pgmap v236: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 5.7 KiB/s rd, 426 B/s wr, 8 op/s 2026-03-09T17:47:45.278 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/662186067' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:45.278 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3662048085' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:45.278 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3654058638' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:45.278 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/4202696512' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:45.278 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1665559662' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:45.278 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1169090237' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:45.278 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3568156390' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:45.278 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/4230412801' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:45.278 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3557055154' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:45.278 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3674915346' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:45.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:45 vm04 ceph-mon[53158]: pgmap v236: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 5.7 KiB/s rd, 426 B/s wr, 8 op/s 2026-03-09T17:47:45.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:45 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/662186067' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:45.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:45 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3662048085' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:45.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:45 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3654058638' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:45.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:45 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/4202696512' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:45.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:45 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1665559662' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:45.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:45 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1169090237' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:45.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:45 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3568156390' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:45.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:45 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/4230412801' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:45.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:45 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3557055154' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:45.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:45 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3674915346' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:45 vm05 ceph-mon[53831]: pgmap v236: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 5.7 KiB/s rd, 426 B/s wr, 8 op/s 2026-03-09T17:47:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:45 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/662186067' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:45 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3662048085' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:45 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3654058638' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:45 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/4202696512' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:45 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1665559662' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:45 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1169090237' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:45 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3568156390' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:45 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/4230412801' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:45 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3557055154' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:45 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3674915346' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug An additional 1 tpg's are required 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Gateway.create_tpg) created TPG '1' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug iqn.2003-01.com.redhat.iscsi-gw:ceph-gw - Adding the IP to the enabled tpg, allowing iSCSI logins 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:45] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:45] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _gateway update on localhost, successful 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:45] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:45] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:45.722 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:45.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.979 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.979 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:45] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:45.979 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:45] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:45.979 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.979 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.979 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.979 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:45] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:45.979 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:45.979 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:45.979 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:45.979 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:45 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:46.290 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:46] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:46.290 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:46.290 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1298113950' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:46.290 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2124473492' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:46.290 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/4056115259' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:46.290 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1780163703' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:46.290 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2754797897' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:46.290 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/4059661479' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:46.290 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3489139381' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:46.290 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/679340383' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:46.290 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3473980840' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:46.290 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3782551996' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:46.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:46 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:46.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:46 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1298113950' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:46.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:46 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2124473492' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:46.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:46 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/4056115259' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:46.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:46 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1780163703' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:46.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:46 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2754797897' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:46.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:46 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/4059661479' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:46.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:46 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3489139381' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:46.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:46 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/679340383' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:46.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:46 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3473980840' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:46.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:46 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3782551996' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:46.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:45 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:47:46.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug Change detected - internal 2 / xattr 3 refreshing 2026-03-09T17:47:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:46 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:46 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1298113950' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:46 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2124473492' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:46 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/4056115259' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:46 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1780163703' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:46 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2754797897' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:46 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/4059661479' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:46 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3489139381' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:46 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/679340383' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:46 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3473980840' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:46.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:46 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3782551996' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:46] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:46] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:46] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:46] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:46] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:46] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:46] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:46] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:46] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug gateway validation needed for vm05.local 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _gateway update on vm05.local, successful 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug An additional 1 tpg's are required 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Gateway.create_tpg) created TPG '2' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-09T17:47:46.554 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _gateway update on vm01.local, successful 2026-03-09T17:47:46.555 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:46] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-09T17:47:46.555 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:46] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-09T17:47:46.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/sysinfo/hostname HTTP/1.1" 200 - 2026-03-09T17:47:46.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/sysinfo/hostname HTTP/1.1" 200 - 2026-03-09T17:47:46.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-09T17:47:46.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-09T17:47:46.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-09T17:47:46.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-09T17:47:46.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-09T17:47:46.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-09T17:47:46.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:46.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:46.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-09T17:47:46.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Gateway.create_tpg) created TPG '1' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-09T17:47:46.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Gateway.create_tpg) created TPG '2' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-09T17:47:46.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Gateway.create_target) created an iscsi target with iqn of 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-09T17:47:46.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug iqn.2003-01.com.redhat.iscsi-gw:ceph-gw - Adding the IP to the enabled tpg, allowing iSCSI logins 2026-03-09T17:47:46.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-09T17:47:46.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-09T17:47:46.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:46.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:46.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:46.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:46] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:47.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:46] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:47.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:46] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:47.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:46 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:46] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:47.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:47] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:47.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:47] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:47.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:47] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:47.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:47] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:47.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-mon[52793]: pgmap v237: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 24 KiB/s rd, 426 B/s wr, 30 op/s 2026-03-09T17:47:47.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:47.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2399628347' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:47.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2768847721' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:47.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3068833782' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:47.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2626534482' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:47.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1359016168' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:47.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3148986287' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:47.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3174813257' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:47.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/455194801' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:47.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3106034200' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:47.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1506036027' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:47.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:47 vm05 ceph-mon[53831]: pgmap v237: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 24 KiB/s rd, 426 B/s wr, 30 op/s 2026-03-09T17:47:47.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:47 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:47.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:47 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2399628347' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:47.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:47 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2768847721' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:47.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:47 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3068833782' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:47.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:47 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2626534482' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:47.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:47 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1359016168' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:47.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:47 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3148986287' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:47.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:47 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3174813257' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:47.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:47 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/455194801' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:47.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:47 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3106034200' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:47.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:47 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1506036027' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:47.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:47 vm04 ceph-mon[53158]: pgmap v237: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 24 KiB/s rd, 426 B/s wr, 30 op/s 2026-03-09T17:47:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:47 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:47 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2399628347' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:47 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2768847721' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:47 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3068833782' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:47 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2626534482' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:47 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1359016168' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:47 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3148986287' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:47 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3174813257' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:47 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/455194801' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:47 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3106034200' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:47.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:47 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1506036027' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:47] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:47] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:47] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:47] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:47] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:47] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:47] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:47] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.683 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block0' to LIO backstore user:rbd 2026-03-09T17:47:47.809 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.810 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.810 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.810 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.810 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.810 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:47.960 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: /usr/lib/python3.9/site-packages/rtslib_fb/root.py:180: UserWarning: Cannot set dbroot to /var/target. Target devices have already been registered. 2026-03-09T17:47:47.960 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: warn("Cannot set dbroot to {}. Target devices have already been registered." 2026-03-09T17:47:47.960 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (LUN.add_dev_to_lio) Successfully added datapool/block0 to LIO 2026-03-09T17:47:47.960 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-09T17:47:47.960 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Setup group ano2 for datapool.block0 on tpg 2 (state 1, owner False, failover type 1) 2026-03-09T17:47:47.960 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Setup group ao for datapool.block0 on tpg 1 (state 0, owner True, failover type 1) 2026-03-09T17:47:47.960 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:47:47.960 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:47:47.960 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _targetlun update on vm01.local, successful 2026-03-09T17:47:47.960 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _targetlun update on vm05.local, successful 2026-03-09T17:47:47.961 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:47] "PUT /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:47:47.961 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:47] "PUT /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:47:47.961 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:47] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:47.961 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:47 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:47] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:48.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:48 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3453105139' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:48.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:48 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1639790661' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:48.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:48 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2867149052' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:48.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:48 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3462998273' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:48.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:48 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/506431081' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:48.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:48 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2403195361' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:47:48.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:48 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2403195361' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:47:48.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:48 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2403195361' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:47:48.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:48 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2403195361' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/3794828077"}]: dispatch 2026-03-09T17:47:48.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:48 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3205701074' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:47:48.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:48 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3205701074' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:47:48.071 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:48 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3205701074' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:47:48.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block0' to LIO backstore user:rbd 2026-03-09T17:47:48.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: /usr/lib/python3.9/site-packages/rtslib_fb/root.py:180: UserWarning: Cannot set dbroot to /var/target. Target devices have already been registered. 2026-03-09T17:47:48.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: warn("Cannot set dbroot to {}. Target devices have already been registered." 2026-03-09T17:47:48.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (LUN.add_dev_to_lio) Successfully added datapool/block0 to LIO 2026-03-09T17:47:48.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (LUN.allocate) added block0 to LIO using wwn '36001405da17b74481464e9fa968746d3' defined by vm01.local 2026-03-09T17:47:48.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-09T17:47:48.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug Setup group ano2 for datapool.block0 on tpg 2 (state 1, owner False, failover type 1) 2026-03-09T17:47:48.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug Setup group ao for datapool.block0 on tpg 1 (state 0, owner True, failover type 1) 2026-03-09T17:47:48.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:47:48.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:47 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:47] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:47:48.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3453105139' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:48.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1639790661' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:48.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2867149052' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:48.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3462998273' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:48.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/506431081' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:48.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2403195361' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:47:48.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2403195361' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:47:48.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2403195361' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:47:48.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2403195361' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/3794828077"}]: dispatch 2026-03-09T17:47:48.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3205701074' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:47:48.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3205701074' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:47:48.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3205701074' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:47:48.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:48 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3453105139' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:48.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:48 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1639790661' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:48.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:48 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2867149052' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:48.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:48 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3462998273' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:48.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:48 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/506431081' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:48.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:48 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2403195361' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:47:48.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:48 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2403195361' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:47:48.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:48 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2403195361' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:47:48.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:48 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2403195361' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/3794828077"}]: dispatch 2026-03-09T17:47:48.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:48 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3205701074' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:47:48.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:48 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3205701074' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:47:48.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:48 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3205701074' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:47:48.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:48.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:48.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:48] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:48.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:48] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:48.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:48.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:48.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:48.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:48.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:48.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:48.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:48.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:48.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:48 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:48.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:48 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:48.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:48 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:48.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:48 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:48] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:48] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Client.define_client) iqn.1994-05.com.redhat:client added successfully 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:48] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:48] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _client update on localhost, successful 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _client update on vm05.local, successful 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:48] "PUT /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:48] "PUT /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:48 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:48] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:49 vm05 ceph-mon[53831]: pgmap v238: 4 pgs: 4 active+clean; 451 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 853 B/s wr, 41 op/s 2026-03-09T17:47:49.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:49 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2821568438' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:49.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:49 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2643050460' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:49.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:49 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1049705791' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:49.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:49 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3400420785' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:49.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:49 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1288165261' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:49.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:49 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2009025058' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:49.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:49 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3107501778' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:49.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:49 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2569386487' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:49.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:49 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3223053956' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:49.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:49 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/353429440' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:49.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:48 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:48 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:48 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:48 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:48 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:48 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:48 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Client.define_client) iqn.1994-05.com.redhat:client added successfully 2026-03-09T17:47:49.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:48 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-09T17:47:49.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:48 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:47:49.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:48 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:48] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:47:49.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-mon[52793]: pgmap v238: 4 pgs: 4 active+clean; 451 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 853 B/s wr, 41 op/s 2026-03-09T17:47:49.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2821568438' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:49.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2643050460' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:49.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1049705791' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:49.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3400420785' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:49.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1288165261' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:49.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2009025058' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:49.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3107501778' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:49.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2569386487' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:49.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3223053956' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:49.418 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/353429440' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:49.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:49 vm04 ceph-mon[53158]: pgmap v238: 4 pgs: 4 active+clean; 451 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 853 B/s wr, 41 op/s 2026-03-09T17:47:49.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:49 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2821568438' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:49.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:49 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2643050460' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:49.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:49 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1049705791' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:49.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:49 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3400420785' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:49.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:49 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1288165261' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:49.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:49 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2009025058' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:49.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:49 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3107501778' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:49.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:49 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2569386487' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:49.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:49 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3223053956' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:49.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:49 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/353429440' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:49.727 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:49] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.727 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:49] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.727 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:49] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:49.727 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:49] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:49.727 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:49] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:49] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:49] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:49] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:49.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:49] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:49] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:49] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:49] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:49 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:49] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:49 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:49] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:49 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:49] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:49.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:49 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:49] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:49] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:50.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:49] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:50.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:49] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:50.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:49 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:49] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:50.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:50] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:50.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:50] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:50.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:50] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:50.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:50] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:50.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/113110652' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:50.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/382174661' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:50.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3678591888' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:50.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/279014858' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:50.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/666782929' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:50.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1346796043' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:50.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3880708303' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:50.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2399075332' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:50.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2216562101' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:50.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/782453240' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:50.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:50 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/113110652' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:50.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:50 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/382174661' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:50.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:50 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3678591888' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:50.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:50 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/279014858' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:50.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:50 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/666782929' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:50.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:50 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1346796043' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:50.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:50 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3880708303' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:50.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:50 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2399075332' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:50.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:50 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2216562101' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:50.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:50 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/782453240' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:50.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:50 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:50 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:50 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:50 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:50 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/113110652' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:50.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:50 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/382174661' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:50.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:50 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3678591888' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:50.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:50 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/279014858' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:50.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:50 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/666782929' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:50.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:50 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1346796043' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:50.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:50 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3880708303' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:50.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:50 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2399075332' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:50.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:50 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2216562101' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:50.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:50 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/782453240' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:50] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:50] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:50] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:50] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:50] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:50] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:50] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:50] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:50] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:50] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.701 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Client.add_lun) added image 'datapool/block0' to iqn.1994-05.com.redhat:client 2026-03-09T17:47:50.701 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-09T17:47:50.701 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:50] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:47:50.701 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:50] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:47:50.701 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _clientlun update on localhost, successful 2026-03-09T17:47:50.701 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _clientlun update on vm05.local, successful 2026-03-09T17:47:50.701 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:50] "PUT /api/clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:47:50.701 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:50] "PUT /api/clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:47:50.701 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:50] "GET /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:47:50.701 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:50 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:50] "GET /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:47:50.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:50 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:50 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:50 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:50 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:50 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:50 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:50.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:50 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Client.add_lun) added image 'datapool/block0' to iqn.1994-05.com.redhat:client 2026-03-09T17:47:50.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:50 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-09T17:47:50.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:50 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:47:50.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:50 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:50] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:47:51.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:51 vm05 ceph-mon[53831]: pgmap v239: 4 pgs: 4 active+clean; 451 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 853 B/s wr, 40 op/s 2026-03-09T17:47:51.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:51 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/123391313' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:51.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:51 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3421507137' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:51.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:51 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3756484230' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:51.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:51 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/690958962' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:51.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:51 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2233798095' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:51.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:51 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1729535018' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:51.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:51 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/2494701829' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:51.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:51 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3048239070' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:51.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:51 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:51.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:51 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:51.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:51 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:51.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:51 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:51.478 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:51] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:51.478 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:51] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:51.478 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:51] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:51.478 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:51] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:51.478 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:51] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:51.478 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:51] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:51.478 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:51] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:51.478 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:51] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:51.478 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:51.478 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:51.478 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:51.478 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-mon[52793]: pgmap v239: 4 pgs: 4 active+clean; 451 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 853 B/s wr, 40 op/s 2026-03-09T17:47:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/123391313' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3421507137' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3756484230' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/690958962' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2233798095' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1729535018' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/2494701829' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:51.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3048239070' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:51 vm04 ceph-mon[53158]: pgmap v239: 4 pgs: 4 active+clean; 451 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 853 B/s wr, 40 op/s 2026-03-09T17:47:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:51 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/123391313' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:51 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3421507137' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:51 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3756484230' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:51 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/690958962' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:51 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2233798095' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:51 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1729535018' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:51 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/2494701829' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:51 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3048239070' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:51.654 INFO:tasks.cram.client.0.vm01.stdout:/home/ubuntu/cephtest/archive/cram.client.0/gwcli_create.t: passed 2026-03-09T17:47:51.654 INFO:tasks.cram.client.0.vm01.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-09T17:47:51.659 INFO:tasks.cram:Running tests for client.1... 2026-03-09T17:47:51.659 DEBUG:teuthology.orchestra.run.vm04:> CEPH_REF=master CEPH_ID="1" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.1/*.t 2026-03-09T17:47:51.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:51] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:51.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:51] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:51.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:51] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:51.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:51] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:47:51.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:51] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:51.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:51] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:51.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:47:51] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:51.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:47:51] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:47:51.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:51.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:51.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:51.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:47:51 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:52.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:51 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:52.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:51 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:52.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:51 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:52.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:47:51 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:47:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:47:52.322 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:52 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3271873081' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:52.322 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:52 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3705580204' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:52.322 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:52 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1606062296' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:52.323 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:52 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/886565037' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:52.323 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:52 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/135344537' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:52.323 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:52 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/1137426635' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:52.323 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:52 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/3974176739' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:52.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:52 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3271873081' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:52.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:52 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3705580204' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:52.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:52 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1606062296' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:52.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:52 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/886565037' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:52.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:52 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/135344537' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:52.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:52 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/1137426635' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:52.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:52 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/3974176739' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:52.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:52 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3271873081' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:52.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:52 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3705580204' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:52.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:52 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1606062296' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:47:52.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:52 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/886565037' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:47:52.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:52 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/135344537' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:52.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:52 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/1137426635' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:47:52.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:52 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/3974176739' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:47:53.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:53 vm01 ceph-mon[52793]: pgmap v240: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 89 KiB/s rd, 2.3 KiB/s wr, 97 op/s 2026-03-09T17:47:53.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:53 vm04 ceph-mon[53158]: pgmap v240: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 89 KiB/s rd, 2.3 KiB/s wr, 97 op/s 2026-03-09T17:47:53.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:53 vm05 ceph-mon[53831]: pgmap v240: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 89 KiB/s rd, 2.3 KiB/s wr, 97 op/s 2026-03-09T17:47:55.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:55 vm01 ceph-mon[52793]: pgmap v241: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 84 KiB/s rd, 1.9 KiB/s wr, 89 op/s 2026-03-09T17:47:55.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:55 vm04 ceph-mon[53158]: pgmap v241: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 84 KiB/s rd, 1.9 KiB/s wr, 89 op/s 2026-03-09T17:47:55.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:55 vm05 ceph-mon[53831]: pgmap v241: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 84 KiB/s rd, 1.9 KiB/s wr, 89 op/s 2026-03-09T17:47:56.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:56 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:56.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:56 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:56.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:56 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:57.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:57 vm01 ceph-mon[52793]: pgmap v242: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 91 KiB/s rd, 1.9 KiB/s wr, 98 op/s 2026-03-09T17:47:57.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:57 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:57.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:57 vm04 ceph-mon[53158]: pgmap v242: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 91 KiB/s rd, 1.9 KiB/s wr, 98 op/s 2026-03-09T17:47:57.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:57 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:57.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:57 vm05 ceph-mon[53831]: pgmap v242: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 91 KiB/s rd, 1.9 KiB/s wr, 98 op/s 2026-03-09T17:47:57.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:57 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:47:59.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:47:59 vm01 ceph-mon[52793]: pgmap v243: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 77 KiB/s rd, 1.9 KiB/s wr, 78 op/s 2026-03-09T17:47:59.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:47:59 vm04 ceph-mon[53158]: pgmap v243: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 77 KiB/s rd, 1.9 KiB/s wr, 78 op/s 2026-03-09T17:47:59.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:47:59 vm05 ceph-mon[53831]: pgmap v243: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 77 KiB/s rd, 1.9 KiB/s wr, 78 op/s 2026-03-09T17:48:01.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:01 vm04 ceph-mon[53158]: pgmap v244: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 1.5 KiB/s wr, 67 op/s 2026-03-09T17:48:01.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:01 vm05 ceph-mon[53831]: pgmap v244: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 1.5 KiB/s wr, 67 op/s 2026-03-09T17:48:01.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:01 vm01 ceph-mon[52793]: pgmap v244: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 1.5 KiB/s wr, 67 op/s 2026-03-09T17:48:02.093 INFO:tasks.cram.client.1.vm04.stdout:/home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t: passed 2026-03-09T17:48:02.093 INFO:tasks.cram.client.1.vm04.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-09T17:48:02.097 INFO:tasks.cram:Running tests for client.2... 2026-03-09T17:48:02.097 DEBUG:teuthology.orchestra.run.vm05:> CEPH_REF=master CEPH_ID="2" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.2/*.t 2026-03-09T17:48:02.539 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:02 vm05 ceph-mon[53831]: pgmap v245: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 68 KiB/s rd, 1.5 KiB/s wr, 68 op/s 2026-03-09T17:48:02.623 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:02 vm01 ceph-mon[52793]: pgmap v245: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 68 KiB/s rd, 1.5 KiB/s wr, 68 op/s 2026-03-09T17:48:02.770 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:02 vm04 ceph-mon[53158]: pgmap v245: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 68 KiB/s rd, 1.5 KiB/s wr, 68 op/s 2026-03-09T17:48:02.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:02] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:02.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:02] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:02.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:02] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:02.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:02] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:02.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:02] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:02.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:02] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:02.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:02] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:02.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:02] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:02.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:02.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:02.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:02.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:02.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:02] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:02.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:02 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:02] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:02.902 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:02 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:02.902 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:02 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:02.902 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:02 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:02.902 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:02 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.176 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.176 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.176 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.176 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.315 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:03] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:03.315 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:03] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:03.315 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:03] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:03.315 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:03] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:03.315 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:03] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:03.315 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:03] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:03.315 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:03] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:03] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:03] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:03] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug _client update on vm01.local, successful 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Client.delete) deleted NodeACL for iqn.1994-05.com.redhat:client 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:03] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:03] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug _client update on localhost, successful 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:03] "DELETE /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:48:03.316 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:03] "DELETE /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:48:03.316 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:03 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/447069977' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:03.316 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:03 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3335833920' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:03.316 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:03 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/412244873' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:03.316 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:03 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1562640422' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:03.316 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:03 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/2640777395' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:03.316 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:03 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/2968971112' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:03.316 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:03 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3082856108' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:03.316 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:03 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3628980291' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:03.316 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:03 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1978324659' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:03.316 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:03 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3087748492' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:03.477 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.477 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.477 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Client.delete) deleted NodeACL for iqn.1994-05.com.redhat:client 2026-03-09T17:48:03.477 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:48:03.477 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:48:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/447069977' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3335833920' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/412244873' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1562640422' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/2640777395' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/2968971112' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3082856108' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3628980291' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1978324659' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:03.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3087748492' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:03.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:03 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/447069977' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:03.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:03 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3335833920' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:03.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:03 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/412244873' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:03.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:03 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1562640422' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:03.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:03 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/2640777395' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:03.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:03 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/2968971112' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:03.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:03 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3082856108' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:03.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:03 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3628980291' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:03.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:03 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1978324659' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:03.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:03 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3087748492' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:03.977 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:03.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:03] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:04.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:03] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:04.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:03] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:04.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:03] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:04.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:03] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:04.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:03] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:04.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:03] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:04.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:03] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:04.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:03 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:03 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Change detected - internal 8 / xattr 9 refreshing 2026-03-09T17:48:04.398 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:04 vm05 ceph-mon[53831]: pgmap v246: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 12 KiB/s rd, 12 op/s 2026-03-09T17:48:04.398 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:04 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/468350395' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:04.398 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:04 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1197417780' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:04.398 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:04 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1585185869' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:04.398 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:04 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1979678632' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:04.398 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:04 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/4144112335' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:04.398 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:04 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/2923521095' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:04.398 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:04 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/4088986791' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:04.398 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:04 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3614213258' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:04.398 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:04 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3374995796' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:04.398 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:04 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/2027943024' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:04] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:04] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:04] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:04] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:04] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:04] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:04] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:04] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug _targetlun update on vm01.local, successful 2026-03-09T17:48:04.398 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug LUN unmap request received, config commit to be performed by vm05.local 2026-03-09T17:48:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:04 vm04 ceph-mon[53158]: pgmap v246: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 12 KiB/s rd, 12 op/s 2026-03-09T17:48:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:04 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/468350395' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:04 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1197417780' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:04 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1585185869' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:04 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1979678632' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:04 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/4144112335' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:04 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/2923521095' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:04 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/4088986791' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:04.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:04 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3614213258' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:04.542 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:04 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3374995796' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:04.542 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:04 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/2027943024' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:04.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-mon[52793]: pgmap v246: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 12 KiB/s rd, 12 op/s 2026-03-09T17:48:04.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/468350395' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:04.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1197417780' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:04.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1585185869' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:04.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1979678632' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:04.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/4144112335' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:04.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/2923521095' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:04.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/4088986791' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:04.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3614213258' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:04.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3374995796' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:04.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/2027943024' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:04.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:04.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug LUN unmap request received, config commit to be performed by vm05.local 2026-03-09T17:48:04.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:48:04.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:48:04.815 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:48:04.815 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:48:04.815 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug _targetlun update on vm05.local, successful 2026-03-09T17:48:04.815 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:04] "DELETE /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:48:04.815 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:04] "DELETE /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:48:05.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:04] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:05.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:04] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:05.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:04] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:05.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:04] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:05.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:04] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:05.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:04] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:05.071 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:04] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:05.072 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:04] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:05.072 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.072 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.072 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.072 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:04 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:48:05.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:04] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Change detected - internal 9 / xattr 10 refreshing 2026-03-09T17:48:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:05 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3570872919' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:05 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3357377624' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:05 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/2137828072' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:05 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:05 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/2295888172' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:05 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3239892349' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:05] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:05] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:05] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:05] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:05] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:05] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:05] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:05] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug _target update on vm01.local, successful 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-09T17:48:05.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug Removing target configuration 2026-03-09T17:48:05.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:05.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-09T17:48:05.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Removing target configuration 2026-03-09T17:48:05.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:48:05.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:48:05.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3570872919' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:05.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3357377624' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:05.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/2137828072' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:05.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:05.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/2295888172' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:05.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3239892349' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:05.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:05 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3570872919' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:05.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:05 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3357377624' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:05.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:05 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/2137828072' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:05.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:05 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:05.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:05 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/2295888172' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:05.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:05 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3239892349' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:05.978 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:48:05.978 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:05] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:48:05.978 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug _target update on vm05.local, successful 2026-03-09T17:48:05.978 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:05] "DELETE /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:48:05.978 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:05] "DELETE /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:48:06.231 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:05] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:06.231 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:05 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:05] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:06.231 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:48:06.231 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:06] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:06.231 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:06] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:06.309 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:05 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Change detected - internal 11 / xattr 12 refreshing 2026-03-09T17:48:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:06 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/2071445292' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:06 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/2147543115' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:06 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3023689194' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:06 vm05 ceph-mon[53831]: pgmap v247: 4 pgs: 4 active+clean; 4.4 MiB data, 216 MiB used, 160 GiB / 160 GiB avail; 83 KiB/s rd, 11 KiB/s wr, 30 op/s 2026-03-09T17:48:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:06 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1659238307' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:06 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3907903057' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:06 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3621456503' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:06 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1682622756' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:06 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/242602371' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:06 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:06.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:06 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3816086067' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:06.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:06 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/4088582694' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:06.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:06 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/2071445292' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:06.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:06 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/2147543115' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:06.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:06 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3023689194' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:06.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:06 vm01 ceph-mon[52793]: pgmap v247: 4 pgs: 4 active+clean; 4.4 MiB data, 216 MiB used, 160 GiB / 160 GiB avail; 83 KiB/s rd, 11 KiB/s wr, 30 op/s 2026-03-09T17:48:06.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:06 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1659238307' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:06.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:06 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3907903057' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:06.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:06 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3621456503' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:06.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:06 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1682622756' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:06.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:06 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/242602371' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:06.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:06 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:06.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:06 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3816086067' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:06.728 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:06 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/4088582694' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:06.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:06 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/2071445292' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:06.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:06 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/2147543115' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:06.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:06 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3023689194' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:06.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:06 vm04 ceph-mon[53158]: pgmap v247: 4 pgs: 4 active+clean; 4.4 MiB data, 216 MiB used, 160 GiB / 160 GiB avail; 83 KiB/s rd, 11 KiB/s wr, 30 op/s 2026-03-09T17:48:06.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:06 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1659238307' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:06.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:06 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3907903057' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:06.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:06 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3621456503' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:06.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:06 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1682622756' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:06.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:06 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/242602371' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:06.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:06 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:06.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:06 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3816086067' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:06.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:06 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/4088582694' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:06.847 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:06] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:06.847 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:06] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:06.847 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:06] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:06.847 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:06] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:06.847 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:06] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:06.847 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:06] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:48:06.847 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug LUN deletion request received, rbd removal to be performed by vm05.local 2026-03-09T17:48:06.847 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:06] "DELETE /api/_disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:06.847 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:06] "DELETE /api/_disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:06.847 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug _disk update on vm05.local, successful 2026-03-09T17:48:06.847 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:06] "DELETE /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:06.847 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:48:06] "DELETE /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-09T17:48:07.290 INFO:tasks.cram.client.2.vm05.stdout:/home/ubuntu/cephtest/archive/cram.client.2/gwcli_delete.t: passed 2026-03-09T17:48:07.290 INFO:tasks.cram.client.2.vm05.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-09T17:48:07.295 DEBUG:teuthology.orchestra.run.vm01:> test -f /home/ubuntu/cephtest/archive/cram.client.0/gwcli_create.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.0/gwcli_create.t 2026-03-09T17:48:07.312 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:07 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1191233070' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:07.312 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:07 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/4143198656' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:07.312 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:07 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1444835889' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:07.312 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:07 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3590338464' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:07.312 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:07 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/935224644' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:07.312 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:07 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3628866198' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:07.312 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:07 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1719603254' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:07.312 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:07 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3525868678' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:07.312 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:07 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/1650877828' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:07.312 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:07 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/2913401465' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:07.312 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:07 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/2484347838' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:07.312 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:06 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Change detected - internal 12 / xattr 13 refreshing 2026-03-09T17:48:07.314 DEBUG:teuthology.orchestra.run.vm01:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.0 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.0 2026-03-09T17:48:07.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:07 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1191233070' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:07.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:07 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/4143198656' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:07.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:07 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1444835889' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:07.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:07 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3590338464' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:07.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:07 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/935224644' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:07.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:07 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3628866198' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:07.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:07 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1719603254' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:07.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:07 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3525868678' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:07.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:07 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/1650877828' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:07.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:07 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/2913401465' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:07.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:07 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/2484347838' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:07.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:07 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:48:07] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:07.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:07 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:48:07] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:48:07.739 DEBUG:teuthology.orchestra.run.vm04:> test -f /home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t 2026-03-09T17:48:07.758 DEBUG:teuthology.orchestra.run.vm04:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.1 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.1 2026-03-09T17:48:07.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:07 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1191233070' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:07.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:07 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/4143198656' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:07.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:07 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1444835889' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:07.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:07 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3590338464' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:07.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:07 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/935224644' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:07.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:07 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3628866198' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:07.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:07 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1719603254' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-09T17:48:07.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:07 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3525868678' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:48:07.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:07 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/1650877828' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:07.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:07 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/2913401465' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T17:48:07.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:07 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/2484347838' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-09T17:48:08.198 DEBUG:teuthology.orchestra.run.vm05:> test -f /home/ubuntu/cephtest/archive/cram.client.2/gwcli_delete.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.2/gwcli_delete.t 2026-03-09T17:48:08.213 DEBUG:teuthology.orchestra.run.vm05:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.2 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.2 2026-03-09T17:48:08.638 INFO:teuthology.run_tasks:Running task cram... 2026-03-09T17:48:08.641 INFO:tasks.cram:Pulling tests from https://github.com/kshtsk/ceph.git ref 569c3e99c9b32a51b4eaf08731c728f4513ed589 2026-03-09T17:48:08.641 DEBUG:teuthology.orchestra.run.vm01:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.0 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-09T17:48:08.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:08 vm01 ceph-mon[52793]: pgmap v248: 4 pgs: 4 active+clean; 11 MiB data, 217 MiB used, 160 GiB / 160 GiB avail; 119 KiB/s rd, 39 KiB/s wr, 44 op/s 2026-03-09T17:48:08.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:08 vm04 ceph-mon[53158]: pgmap v248: 4 pgs: 4 active+clean; 11 MiB data, 217 MiB used, 160 GiB / 160 GiB avail; 119 KiB/s rd, 39 KiB/s wr, 44 op/s 2026-03-09T17:48:08.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:08 vm05 ceph-mon[53831]: pgmap v248: 4 pgs: 4 active+clean; 11 MiB data, 217 MiB used, 160 GiB / 160 GiB avail; 119 KiB/s rd, 39 KiB/s wr, 44 op/s 2026-03-09T17:48:10.186 INFO:teuthology.orchestra.run.vm01.stdout:Collecting cram==0.6 2026-03-09T17:48:10.187 INFO:teuthology.orchestra.run.vm01.stdout: Using cached cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-09T17:48:10.199 INFO:teuthology.orchestra.run.vm01.stdout:Installing collected packages: cram 2026-03-09T17:48:10.206 INFO:teuthology.orchestra.run.vm01.stdout:Successfully installed cram-0.6 2026-03-09T17:48:10.207 INFO:teuthology.orchestra.run.vm01.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-03-09T17:48:10.207 INFO:teuthology.orchestra.run.vm01.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/virtualenv/bin/python3 -m pip install --upgrade pip' command. 2026-03-09T17:48:10.251 DEBUG:teuthology.orchestra.run.vm01:> rm -rf /home/ubuntu/cephtest/clone.client.0 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.0 && cd /home/ubuntu/cephtest/clone.client.0 && git checkout 569c3e99c9b32a51b4eaf08731c728f4513ed589 2026-03-09T17:48:10.270 INFO:teuthology.orchestra.run.vm01.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.0'... 2026-03-09T17:48:10.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:10 vm04 ceph-mon[53158]: pgmap v249: 4 pgs: 4 active+clean; 11 MiB data, 217 MiB used, 160 GiB / 160 GiB avail; 115 KiB/s rd, 39 KiB/s wr, 42 op/s 2026-03-09T17:48:10.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:10 vm05 ceph-mon[53831]: pgmap v249: 4 pgs: 4 active+clean; 11 MiB data, 217 MiB used, 160 GiB / 160 GiB avail; 115 KiB/s rd, 39 KiB/s wr, 42 op/s 2026-03-09T17:48:10.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:10 vm01 ceph-mon[52793]: pgmap v249: 4 pgs: 4 active+clean; 11 MiB data, 217 MiB used, 160 GiB / 160 GiB avail; 115 KiB/s rd, 39 KiB/s wr, 42 op/s 2026-03-09T17:48:12.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:12 vm04 ceph-mon[53158]: pgmap v250: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 134 KiB/s rd, 39 KiB/s wr, 67 op/s 2026-03-09T17:48:12.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:12 vm05 ceph-mon[53831]: pgmap v250: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 134 KiB/s rd, 39 KiB/s wr, 67 op/s 2026-03-09T17:48:12.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:12 vm01 ceph-mon[52793]: pgmap v250: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 134 KiB/s rd, 39 KiB/s wr, 67 op/s 2026-03-09T17:48:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:14 vm04 ceph-mon[53158]: pgmap v251: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 66 op/s 2026-03-09T17:48:14.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:14 vm05 ceph-mon[53831]: pgmap v251: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 66 op/s 2026-03-09T17:48:14.821 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:14 vm01 ceph-mon[52793]: pgmap v251: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 66 op/s 2026-03-09T17:48:14.821 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:14 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:48:15.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:15 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:15.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:15 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:15.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:15 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:16.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:48:16.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:16 vm05 ceph-mon[53831]: pgmap v252: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 66 op/s 2026-03-09T17:48:16.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:16 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:16.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:16 vm01 ceph-mon[52793]: pgmap v252: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 66 op/s 2026-03-09T17:48:16.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:16 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:17.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:16 vm04 ceph-mon[53158]: pgmap v252: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 66 op/s 2026-03-09T17:48:17.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:16 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:18.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:18 vm05 ceph-mon[53831]: pgmap v253: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 62 KiB/s rd, 28 KiB/s wr, 49 op/s 2026-03-09T17:48:18.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:18 vm01 ceph-mon[52793]: pgmap v253: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 62 KiB/s rd, 28 KiB/s wr, 49 op/s 2026-03-09T17:48:19.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:18 vm04 ceph-mon[53158]: pgmap v253: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 62 KiB/s rd, 28 KiB/s wr, 49 op/s 2026-03-09T17:48:20.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:20 vm05 ceph-mon[53831]: pgmap v254: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 20 KiB/s rd, 682 B/s wr, 26 op/s 2026-03-09T17:48:20.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:20 vm01 ceph-mon[52793]: pgmap v254: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 20 KiB/s rd, 682 B/s wr, 26 op/s 2026-03-09T17:48:21.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:20 vm04 ceph-mon[53158]: pgmap v254: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 20 KiB/s rd, 682 B/s wr, 26 op/s 2026-03-09T17:48:22.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:22 vm05 ceph-mon[53831]: pgmap v255: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 21 KiB/s rd, 682 B/s wr, 26 op/s 2026-03-09T17:48:22.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:22 vm01 ceph-mon[52793]: pgmap v255: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 21 KiB/s rd, 682 B/s wr, 26 op/s 2026-03-09T17:48:23.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:22 vm04 ceph-mon[53158]: pgmap v255: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 21 KiB/s rd, 682 B/s wr, 26 op/s 2026-03-09T17:48:24.830 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:24 vm01 ceph-mon[52793]: pgmap v256: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:24.843 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:24 vm04 ceph-mon[53158]: pgmap v256: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:25.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:24 vm05 ceph-mon[53831]: pgmap v256: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:25.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:24 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:48:25.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:25 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:26.022 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:25 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:26.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:25 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:26.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:26 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:48:26.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:26 vm01 ceph-mon[52793]: pgmap v257: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:26.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:26 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:27.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:26 vm04 ceph-mon[53158]: pgmap v257: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:27.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:26 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:27.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:26 vm05 ceph-mon[53831]: pgmap v257: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:27.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:26 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:28.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:28 vm01 ceph-mon[52793]: pgmap v258: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:29.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:28 vm04 ceph-mon[53158]: pgmap v258: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:29.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:28 vm05 ceph-mon[53831]: pgmap v258: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:30.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:30 vm01 ceph-mon[52793]: pgmap v259: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:31.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:30 vm04 ceph-mon[53158]: pgmap v259: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:31.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:30 vm05 ceph-mon[53831]: pgmap v259: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:32.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:32 vm01 ceph-mon[52793]: pgmap v260: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:33.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:32 vm04 ceph-mon[53158]: pgmap v260: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:33.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:32 vm05 ceph-mon[53831]: pgmap v260: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:34.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:34 vm01 ceph-mon[52793]: pgmap v261: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:34.978 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:34 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:48:35.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:34 vm04 ceph-mon[53158]: pgmap v261: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:35.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:34 vm05 ceph-mon[53831]: pgmap v261: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:36 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:36.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:36 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:36.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:36 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:48:36.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:35 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:37 vm04 ceph-mon[53158]: pgmap v262: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:37 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:37.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:37 vm05 ceph-mon[53831]: pgmap v262: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:37.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:37 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:37.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:37 vm01 ceph-mon[52793]: pgmap v262: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:37.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:37 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:39 vm04 ceph-mon[53158]: pgmap v263: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:39 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:48:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:39 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:48:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:39 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:48:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:39 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:48:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:39 vm05 ceph-mon[53831]: pgmap v263: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:39 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:48:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:39 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:48:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:39 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:48:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:39 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:48:39.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:39 vm01 ceph-mon[52793]: pgmap v263: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:39.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:39 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:48:39.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:39 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:48:39.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:39 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:48:39.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:39 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:48:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:41 vm04 ceph-mon[53158]: pgmap v264: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:41 vm05 ceph-mon[53831]: pgmap v264: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:41.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:41 vm01 ceph-mon[52793]: pgmap v264: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:43 vm04 ceph-mon[53158]: pgmap v265: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:43 vm05 ceph-mon[53831]: pgmap v265: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:43.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:43 vm01 ceph-mon[52793]: pgmap v265: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:45.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:48:45.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:45 vm01 ceph-mon[52793]: pgmap v266: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:45.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:45 vm04 ceph-mon[53158]: pgmap v266: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:45 vm05 ceph-mon[53831]: pgmap v266: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:46.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:46 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:46.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:46 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:46.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:48:46.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:46 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:47.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:47 vm04 ceph-mon[53158]: pgmap v267: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:47.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:47 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:47.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:47 vm05 ceph-mon[53831]: pgmap v267: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:47.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:47 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:47.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:47 vm01 ceph-mon[52793]: pgmap v267: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:47.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:47 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:49.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:49 vm05 ceph-mon[53831]: pgmap v268: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:49.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:49 vm01 ceph-mon[52793]: pgmap v268: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:49.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:49 vm04 ceph-mon[53158]: pgmap v268: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:51.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:51 vm05 ceph-mon[53831]: pgmap v269: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:51.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:51 vm01 ceph-mon[52793]: pgmap v269: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:51.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:51 vm04 ceph-mon[53158]: pgmap v269: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:53.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:53 vm05 ceph-mon[53831]: pgmap v270: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:53.451 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:53 vm01 ceph-mon[52793]: pgmap v270: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:53.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:53 vm04 ceph-mon[53158]: pgmap v270: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:55.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:55 vm01 ceph-mon[52793]: pgmap v271: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:55.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:48:54 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:48:55.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:55 vm05 ceph-mon[53831]: pgmap v271: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:55.343 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:55 vm04 ceph-mon[53158]: pgmap v271: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:56.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:56 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:56.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:48:56 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:48:56.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:56 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:56.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:56 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:57.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:57 vm05 ceph-mon[53831]: pgmap v272: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:57.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:57 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:57.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:57 vm01 ceph-mon[52793]: pgmap v272: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:57.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:57 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:57.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:57 vm04 ceph-mon[53158]: pgmap v272: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:48:57.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:57 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:48:59.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:48:59 vm01 ceph-mon[52793]: pgmap v273: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:59.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:48:59 vm04 ceph-mon[53158]: pgmap v273: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:48:59.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:48:59 vm05 ceph-mon[53831]: pgmap v273: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:01.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:01 vm01 ceph-mon[52793]: pgmap v274: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:01.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:01 vm04 ceph-mon[53158]: pgmap v274: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:01.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:01 vm05 ceph-mon[53831]: pgmap v274: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:03.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:03 vm01 ceph-mon[52793]: pgmap v275: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:03.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:03 vm04 ceph-mon[53158]: pgmap v275: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:03.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:03 vm05 ceph-mon[53831]: pgmap v275: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:05.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:05 vm01 ceph-mon[52793]: pgmap v276: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:05.228 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:49:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:49:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:05 vm04 ceph-mon[53158]: pgmap v276: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:05.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:05 vm05 ceph-mon[53831]: pgmap v276: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:06.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:06 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:06.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:49:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:49:06.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:06 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:06.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:06 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:07.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:07 vm01 ceph-mon[52793]: pgmap v277: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:07.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:07 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:07.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:07 vm04 ceph-mon[53158]: pgmap v277: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:07.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:07 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:07.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:07 vm05 ceph-mon[53831]: pgmap v277: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:07.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:07 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:08.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:08 vm01 ceph-mon[52793]: pgmap v278: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:08.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:08 vm04 ceph-mon[53158]: pgmap v278: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:08.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:08 vm05 ceph-mon[53831]: pgmap v278: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:10.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:10 vm04 ceph-mon[53158]: pgmap v279: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:10.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:10 vm05 ceph-mon[53831]: pgmap v279: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:10.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:10 vm01 ceph-mon[52793]: pgmap v279: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:12.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:12 vm04 ceph-mon[53158]: pgmap v280: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:12.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:12 vm05 ceph-mon[53831]: pgmap v280: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:12.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:12 vm01 ceph-mon[52793]: pgmap v280: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:14.397 INFO:teuthology.orchestra.run.vm01.stderr:Note: switching to '569c3e99c9b32a51b4eaf08731c728f4513ed589'. 2026-03-09T17:49:14.397 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:49:14.397 INFO:teuthology.orchestra.run.vm01.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-09T17:49:14.397 INFO:teuthology.orchestra.run.vm01.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-09T17:49:14.398 INFO:teuthology.orchestra.run.vm01.stderr:state without impacting any branches by switching back to a branch. 2026-03-09T17:49:14.398 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:49:14.398 INFO:teuthology.orchestra.run.vm01.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-09T17:49:14.398 INFO:teuthology.orchestra.run.vm01.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-09T17:49:14.398 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:49:14.398 INFO:teuthology.orchestra.run.vm01.stderr: git switch -c 2026-03-09T17:49:14.398 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:49:14.398 INFO:teuthology.orchestra.run.vm01.stderr:Or undo this operation with: 2026-03-09T17:49:14.398 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:49:14.398 INFO:teuthology.orchestra.run.vm01.stderr: git switch - 2026-03-09T17:49:14.398 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:49:14.398 INFO:teuthology.orchestra.run.vm01.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-09T17:49:14.398 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:49:14.398 INFO:teuthology.orchestra.run.vm01.stderr:HEAD is now at 569c3e99c9b qa/rgw: bucket notifications use pynose 2026-03-09T17:49:14.403 DEBUG:teuthology.orchestra.run.vm01:> cp -- /home/ubuntu/cephtest/clone.client.0/src/test/cli-integration/rbd/rest_api_create.t /home/ubuntu/cephtest/archive/cram.client.0 2026-03-09T17:49:14.461 DEBUG:teuthology.orchestra.run.vm04:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.1 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-09T17:49:14.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:14 vm05 ceph-mon[53831]: pgmap v281: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:14.883 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:14 vm01 ceph-mon[52793]: pgmap v281: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:15.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:14 vm04 ceph-mon[53158]: pgmap v281: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:15.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:49:14 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:49:15.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:15 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:15.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:15 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:15.993 INFO:teuthology.orchestra.run.vm04.stdout:Collecting cram==0.6 2026-03-09T17:49:15.994 INFO:teuthology.orchestra.run.vm04.stdout: Using cached cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-09T17:49:16.008 INFO:teuthology.orchestra.run.vm04.stdout:Installing collected packages: cram 2026-03-09T17:49:16.016 INFO:teuthology.orchestra.run.vm04.stdout:Successfully installed cram-0.6 2026-03-09T17:49:16.018 INFO:teuthology.orchestra.run.vm04.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-03-09T17:49:16.018 INFO:teuthology.orchestra.run.vm04.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/virtualenv/bin/python3 -m pip install --upgrade pip' command. 2026-03-09T17:49:16.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:15 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:16.061 DEBUG:teuthology.orchestra.run.vm04:> rm -rf /home/ubuntu/cephtest/clone.client.1 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.1 && cd /home/ubuntu/cephtest/clone.client.1 && git checkout 569c3e99c9b32a51b4eaf08731c728f4513ed589 2026-03-09T17:49:16.078 INFO:teuthology.orchestra.run.vm04.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.1'... 2026-03-09T17:49:16.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:49:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:49:16.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:16 vm05 ceph-mon[53831]: pgmap v282: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:16.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:16 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:16.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:16 vm01 ceph-mon[52793]: pgmap v282: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:16.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:16 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:17.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:16 vm04 ceph-mon[53158]: pgmap v282: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:17.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:16 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:18.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:18 vm01 ceph-mon[52793]: pgmap v283: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:19.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:18 vm04 ceph-mon[53158]: pgmap v283: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:19.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:18 vm05 ceph-mon[53831]: pgmap v283: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:20.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:20 vm01 ceph-mon[52793]: pgmap v284: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:21.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:20 vm04 ceph-mon[53158]: pgmap v284: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:21.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:20 vm05 ceph-mon[53831]: pgmap v284: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:22.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:22 vm01 ceph-mon[52793]: pgmap v285: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:23.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:22 vm04 ceph-mon[53158]: pgmap v285: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:23.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:22 vm05 ceph-mon[53831]: pgmap v285: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:24.872 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:24 vm05 ceph-mon[53831]: pgmap v286: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:24.889 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:24 vm01 ceph-mon[52793]: pgmap v286: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:25.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:24 vm04 ceph-mon[53158]: pgmap v286: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:25.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:49:24 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:49:25.942 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:25 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:25.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:25 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:26.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:25 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:26.570 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:49:26 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:49:26.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:26 vm01 ceph-mon[52793]: pgmap v287: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:26.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:26 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:27.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:26 vm04 ceph-mon[53158]: pgmap v287: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:27.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:26 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:27.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:26 vm05 ceph-mon[53831]: pgmap v287: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:27.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:26 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:28.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:28 vm01 ceph-mon[52793]: pgmap v288: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:29.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:28 vm04 ceph-mon[53158]: pgmap v288: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:29.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:28 vm05 ceph-mon[53831]: pgmap v288: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:31.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:30 vm04 ceph-mon[53158]: pgmap v289: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:31.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:30 vm05 ceph-mon[53831]: pgmap v289: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:31.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:30 vm01 ceph-mon[52793]: pgmap v289: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:33.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:33 vm04 ceph-mon[53158]: pgmap v290: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:33.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:33 vm05 ceph-mon[53831]: pgmap v290: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:33.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:33 vm01 ceph-mon[52793]: pgmap v290: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:35.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:49:34 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:49:35.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:35 vm01 ceph-mon[52793]: pgmap v291: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:35 vm04 ceph-mon[53158]: pgmap v291: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:35.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:35 vm05 ceph-mon[53831]: pgmap v291: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:36 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:36.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:36 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:36.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:49:36 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:49:36.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:36 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:37 vm04 ceph-mon[53158]: pgmap v292: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:37 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:37.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:37 vm05 ceph-mon[53831]: pgmap v292: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:37.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:37 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:37.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:37 vm01 ceph-mon[52793]: pgmap v292: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:37.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:37 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:39 vm04 ceph-mon[53158]: pgmap v293: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:39 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:49:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:39 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:49:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:39 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:49:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:39 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:49:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:39 vm05 ceph-mon[53831]: pgmap v293: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:39 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:49:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:39 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:49:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:39 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:49:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:39 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:49:39.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:39 vm01 ceph-mon[52793]: pgmap v293: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:39.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:39 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:49:39.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:39 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:49:39.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:39 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:49:39.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:39 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:49:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:41 vm04 ceph-mon[53158]: pgmap v294: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:41 vm05 ceph-mon[53831]: pgmap v294: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:41.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:41 vm01 ceph-mon[52793]: pgmap v294: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:42.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:42 vm04 ceph-mon[53158]: pgmap v295: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:42.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:42 vm05 ceph-mon[53831]: pgmap v295: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:42.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:42 vm01 ceph-mon[52793]: pgmap v295: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:44.977 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:49:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:49:44.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:44 vm01 ceph-mon[52793]: pgmap v296: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:45.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:44 vm04 ceph-mon[53158]: pgmap v296: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:45.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:44 vm05 ceph-mon[53831]: pgmap v296: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:45.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:45 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:46.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:45 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:46.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:45 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:46.570 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:49:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:49:46.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:46 vm01 ceph-mon[52793]: pgmap v297: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:46.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:46 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:47.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:46 vm04 ceph-mon[53158]: pgmap v297: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:47.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:46 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:47.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:46 vm05 ceph-mon[53831]: pgmap v297: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:47.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:46 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:49.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:48 vm01 ceph-mon[52793]: pgmap v298: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:49.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:48 vm04 ceph-mon[53158]: pgmap v298: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:49.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:48 vm05 ceph-mon[53831]: pgmap v298: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:51.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:50 vm01 ceph-mon[52793]: pgmap v299: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:51.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:50 vm04 ceph-mon[53158]: pgmap v299: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:51.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:50 vm05 ceph-mon[53831]: pgmap v299: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:53.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:52 vm01 ceph-mon[52793]: pgmap v300: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:53.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:52 vm04 ceph-mon[53158]: pgmap v300: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:53.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:52 vm05 ceph-mon[53831]: pgmap v300: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:55.123 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:54 vm05 ceph-mon[53831]: pgmap v301: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:55.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:49:54 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:49:55.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:54 vm01 ceph-mon[52793]: pgmap v301: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:55.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:54 vm04 ceph-mon[53158]: pgmap v301: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:56.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:55 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:56.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:55 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:56.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:55 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:56.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:49:56 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:49:57.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:56 vm01 ceph-mon[52793]: pgmap v302: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:57.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:56 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:57.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:56 vm04 ceph-mon[53158]: pgmap v302: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:57.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:56 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:57.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:56 vm05 ceph-mon[53831]: pgmap v302: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:49:57.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:56 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:49:59.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:49:58 vm01 ceph-mon[52793]: pgmap v303: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:49:58 vm04 ceph-mon[53158]: pgmap v303: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:49:59.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:49:58 vm05 ceph-mon[53831]: pgmap v303: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:01.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:01 vm04 ceph-mon[53158]: pgmap v304: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:01.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:01 vm04 ceph-mon[53158]: overall HEALTH_OK 2026-03-09T17:50:01.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:01 vm05 ceph-mon[53831]: pgmap v304: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:01.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:01 vm05 ceph-mon[53831]: overall HEALTH_OK 2026-03-09T17:50:01.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:01 vm01 ceph-mon[52793]: pgmap v304: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:01.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:01 vm01 ceph-mon[52793]: overall HEALTH_OK 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr:Note: switching to '569c3e99c9b32a51b4eaf08731c728f4513ed589'. 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr:state without impacting any branches by switching back to a branch. 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr: git switch -c 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr:Or undo this operation with: 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr: git switch - 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:50:02.862 INFO:teuthology.orchestra.run.vm04.stderr:HEAD is now at 569c3e99c9b qa/rgw: bucket notifications use pynose 2026-03-09T17:50:02.868 DEBUG:teuthology.orchestra.run.vm04:> cp -- /home/ubuntu/cephtest/clone.client.1/src/test/cli-integration/rbd/iscsi_client.t /home/ubuntu/cephtest/archive/cram.client.1 2026-03-09T17:50:02.927 DEBUG:teuthology.orchestra.run.vm05:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.2 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-09T17:50:03.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:03 vm05 ceph-mon[53831]: pgmap v305: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:03.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:03 vm01 ceph-mon[52793]: pgmap v305: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:03.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:03 vm04 ceph-mon[53158]: pgmap v305: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:04.447 INFO:teuthology.orchestra.run.vm05.stdout:Collecting cram==0.6 2026-03-09T17:50:04.448 INFO:teuthology.orchestra.run.vm05.stdout: Using cached cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-09T17:50:04.458 INFO:teuthology.orchestra.run.vm05.stdout:Installing collected packages: cram 2026-03-09T17:50:04.464 INFO:teuthology.orchestra.run.vm05.stdout:Successfully installed cram-0.6 2026-03-09T17:50:04.465 INFO:teuthology.orchestra.run.vm05.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-03-09T17:50:04.465 INFO:teuthology.orchestra.run.vm05.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/virtualenv/bin/python3 -m pip install --upgrade pip' command. 2026-03-09T17:50:04.499 DEBUG:teuthology.orchestra.run.vm05:> rm -rf /home/ubuntu/cephtest/clone.client.2 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.2 && cd /home/ubuntu/cephtest/clone.client.2 && git checkout 569c3e99c9b32a51b4eaf08731c728f4513ed589 2026-03-09T17:50:04.515 INFO:teuthology.orchestra.run.vm05.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.2'... 2026-03-09T17:50:05.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:50:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:50:05.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:05 vm01 ceph-mon[52793]: pgmap v306: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:05.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:05 vm05 ceph-mon[53831]: pgmap v306: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:05 vm04 ceph-mon[53158]: pgmap v306: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:06.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:06 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:06.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:06 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:06 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:06.570 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:50:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:50:07.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:07 vm01 ceph-mon[52793]: pgmap v307: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:07.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:07 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:07.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:07 vm04 ceph-mon[53158]: pgmap v307: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:07.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:07 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:07.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:07 vm05 ceph-mon[53831]: pgmap v307: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:07.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:07 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:09.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:09 vm01 ceph-mon[52793]: pgmap v308: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:09.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:09 vm04 ceph-mon[53158]: pgmap v308: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:09.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:09 vm05 ceph-mon[53831]: pgmap v308: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:11.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:11 vm01 ceph-mon[52793]: pgmap v309: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:11.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:11 vm04 ceph-mon[53158]: pgmap v309: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:11.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:11 vm05 ceph-mon[53831]: pgmap v309: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:13.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:13 vm05 ceph-mon[53831]: pgmap v310: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:13.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:13 vm01 ceph-mon[52793]: pgmap v310: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:13.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:13 vm04 ceph-mon[53158]: pgmap v310: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:14.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:14 vm01 ceph-mon[52793]: pgmap v311: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:14.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:14 vm04 ceph-mon[53158]: pgmap v311: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:14.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:14 vm05 ceph-mon[53831]: pgmap v311: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:15.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:50:14 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:50:15.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:15 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:15.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:15 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:15.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:15 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:16.506 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:16 vm05 ceph-mon[53831]: pgmap v312: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:16.507 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:16 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:16.507 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:50:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:50:16.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:16 vm04 ceph-mon[53158]: pgmap v312: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:16.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:16 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:16.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:16 vm01 ceph-mon[52793]: pgmap v312: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:16.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:16 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:18.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:18 vm01 ceph-mon[52793]: pgmap v313: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:19.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:18 vm04 ceph-mon[53158]: pgmap v313: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:19.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:18 vm05 ceph-mon[53831]: pgmap v313: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:21.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:20 vm04 ceph-mon[53158]: pgmap v314: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:21.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:20 vm05 ceph-mon[53831]: pgmap v314: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:21.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:20 vm01 ceph-mon[52793]: pgmap v314: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:23.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:22 vm05 ceph-mon[53831]: pgmap v315: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:23.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:22 vm01 ceph-mon[52793]: pgmap v315: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:23.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:22 vm04 ceph-mon[53158]: pgmap v315: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:25.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:24 vm05 ceph-mon[53831]: pgmap v316: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:25.093 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:24 vm04 ceph-mon[53158]: pgmap v316: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:25.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:50:24 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:50:25.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:24 vm01 ceph-mon[52793]: pgmap v316: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:26.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:25 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:26.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:25 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:26.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:25 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:26.570 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:50:26 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:50:27.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:27 vm04 ceph-mon[53158]: pgmap v317: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:27.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:27 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:27.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:27 vm05 ceph-mon[53831]: pgmap v317: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:27.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:27 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:27.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:27 vm01 ceph-mon[52793]: pgmap v317: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:27.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:27 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:29.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:29 vm04 ceph-mon[53158]: pgmap v318: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:29.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:29 vm05 ceph-mon[53831]: pgmap v318: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:29.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:29 vm01 ceph-mon[52793]: pgmap v318: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:31.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:31 vm05 ceph-mon[53831]: pgmap v319: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:31.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:31 vm01 ceph-mon[52793]: pgmap v319: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:31.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:31 vm04 ceph-mon[53158]: pgmap v319: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:33.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:33 vm01 ceph-mon[52793]: pgmap v320: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:33.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:33 vm04 ceph-mon[53158]: pgmap v320: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:33.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:33 vm05 ceph-mon[53831]: pgmap v320: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:34.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:34 vm04 ceph-mon[53158]: pgmap v321: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:34.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:34 vm05 ceph-mon[53831]: pgmap v321: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:34.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:34 vm01 ceph-mon[52793]: pgmap v321: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:35.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:50:34 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:50:35.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:35 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:35.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:35 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:35.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:35 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:36.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:36 vm04 ceph-mon[53158]: pgmap v322: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:36.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:36 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:36.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:36 vm05 ceph-mon[53831]: pgmap v322: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:36.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:36 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:36.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:50:36 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:50:36.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:36 vm01 ceph-mon[52793]: pgmap v322: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:36.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:36 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:38.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:38 vm04 ceph-mon[53158]: pgmap v323: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:38.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:38 vm05 ceph-mon[53831]: pgmap v323: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:38.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:38 vm01 ceph-mon[52793]: pgmap v323: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:39.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:39 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:50:39.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:39 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:50:39.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:39 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:50:39.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:39 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:50:39.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:39 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:50:39.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:39 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:50:39.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:39 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:50:39.821 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:39 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:50:39.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:39 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:50:39.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:39 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:50:39.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:39 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:50:39.978 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:39 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:50:40.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:40 vm04 ceph-mon[53158]: pgmap v324: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:40.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:40 vm05 ceph-mon[53831]: pgmap v324: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:40.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:40 vm01 ceph-mon[52793]: pgmap v324: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:42.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:42 vm04 ceph-mon[53158]: pgmap v325: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:42.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:42 vm05 ceph-mon[53831]: pgmap v325: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:42.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:42 vm01 ceph-mon[52793]: pgmap v325: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:44.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:44 vm04 ceph-mon[53158]: pgmap v326: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:44.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:44 vm05 ceph-mon[53831]: pgmap v326: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:44.939 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:44 vm01 ceph-mon[52793]: pgmap v326: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:45.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:50:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:50:45.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:45 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:45.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:45 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:45.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:45 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:46.537 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:50:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:50:46.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:46 vm04 ceph-mon[53158]: pgmap v327: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:46.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:46 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:46.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:46 vm05 ceph-mon[53831]: pgmap v327: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:46.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:46 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:46.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:46 vm01 ceph-mon[52793]: pgmap v327: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:46.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:46 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:48.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:48 vm05 ceph-mon[53831]: pgmap v328: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:48.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:48 vm01 ceph-mon[52793]: pgmap v328: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:49.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:48 vm04 ceph-mon[53158]: pgmap v328: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:50.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:50 vm05 ceph-mon[53831]: pgmap v329: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:50.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:50 vm01 ceph-mon[52793]: pgmap v329: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:51.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:50 vm04 ceph-mon[53158]: pgmap v329: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:52.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:52 vm01 ceph-mon[52793]: pgmap v330: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:53.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:52 vm04 ceph-mon[53158]: pgmap v330: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:53.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:52 vm05 ceph-mon[53831]: pgmap v330: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:54.945 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:54 vm01 ceph-mon[52793]: pgmap v331: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:55.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:54 vm04 ceph-mon[53158]: pgmap v331: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:55.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:54 vm05 ceph-mon[53831]: pgmap v331: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:55.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:50:54 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:50:55.942 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:55 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:55.972 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:55 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:55.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:55 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:56.240 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:50:56 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:50:57.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:56 vm04 ceph-mon[53158]: pgmap v332: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:57.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:56 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:57.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:56 vm05 ceph-mon[53831]: pgmap v332: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:57.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:56 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:57.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:56 vm01 ceph-mon[52793]: pgmap v332: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:50:57.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:56 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:50:59.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:50:58 vm05 ceph-mon[53831]: pgmap v333: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:59.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:50:58 vm01 ceph-mon[52793]: pgmap v333: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:50:59.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:50:58 vm04 ceph-mon[53158]: pgmap v333: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:01.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:00 vm05 ceph-mon[53831]: pgmap v334: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:01.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:00 vm01 ceph-mon[52793]: pgmap v334: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:01.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:00 vm04 ceph-mon[53158]: pgmap v334: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:03.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:02 vm05 ceph-mon[53831]: pgmap v335: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:03.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:02 vm01 ceph-mon[52793]: pgmap v335: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:03.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:02 vm04 ceph-mon[53158]: pgmap v335: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:05.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:04 vm05 ceph-mon[53831]: pgmap v336: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:05.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:51:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:51:05.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:04 vm01 ceph-mon[52793]: pgmap v336: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:05.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:04 vm04 ceph-mon[53158]: pgmap v336: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:06.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:05 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:06.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:05 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:06.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:05 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:06.570 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:51:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:51:07.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:06 vm05 ceph-mon[53831]: pgmap v337: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:07.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:06 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:07.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:06 vm01 ceph-mon[52793]: pgmap v337: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:07.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:06 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:07.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:06 vm04 ceph-mon[53158]: pgmap v337: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:07.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:06 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:09.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:08 vm01 ceph-mon[52793]: pgmap v338: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:09.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:08 vm04 ceph-mon[53158]: pgmap v338: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:09.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:08 vm05 ceph-mon[53831]: pgmap v338: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:11.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:10 vm01 ceph-mon[52793]: pgmap v339: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:11.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:10 vm04 ceph-mon[53158]: pgmap v339: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:11.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:10 vm05 ceph-mon[53831]: pgmap v339: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:13.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:12 vm01 ceph-mon[52793]: pgmap v340: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:13.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:12 vm04 ceph-mon[53158]: pgmap v340: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:13.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:12 vm05 ceph-mon[53831]: pgmap v340: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:15.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:51:14 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:51:15.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:14 vm01 ceph-mon[52793]: pgmap v341: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:15.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:14 vm04 ceph-mon[53158]: pgmap v341: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:15.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:14 vm05 ceph-mon[53831]: pgmap v341: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:16.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:15 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:16.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:51:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:51:16.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:15 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:16.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:15 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:17.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:16 vm04 ceph-mon[53158]: pgmap v342: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:17.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:16 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:17.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:16 vm05 ceph-mon[53831]: pgmap v342: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:17.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:16 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:17.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:16 vm01 ceph-mon[52793]: pgmap v342: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:17.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:16 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:19.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:18 vm04 ceph-mon[53158]: pgmap v343: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:19.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:18 vm05 ceph-mon[53831]: pgmap v343: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:19.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:18 vm01 ceph-mon[52793]: pgmap v343: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:21.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:20 vm04 ceph-mon[53158]: pgmap v344: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:21.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:20 vm05 ceph-mon[53831]: pgmap v344: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:21.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:20 vm01 ceph-mon[52793]: pgmap v344: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:23.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:23 vm04 ceph-mon[53158]: pgmap v345: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:23.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:23 vm05 ceph-mon[53831]: pgmap v345: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:23.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:23 vm01 ceph-mon[52793]: pgmap v345: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:25.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:51:24 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:51:25.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:25 vm01 ceph-mon[52793]: pgmap v346: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:25.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:25 vm04 ceph-mon[53158]: pgmap v346: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:25.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:25 vm05 ceph-mon[53831]: pgmap v346: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:26.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:26 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:26.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:26 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:26.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:51:26 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:51:26.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:26 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:27.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:27 vm04 ceph-mon[53158]: pgmap v347: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:27.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:27 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:27.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:27 vm05 ceph-mon[53831]: pgmap v347: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:27.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:27 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:27.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:27 vm01 ceph-mon[52793]: pgmap v347: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:27.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:27 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:29.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:29 vm04 ceph-mon[53158]: pgmap v348: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:29.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:29 vm05 ceph-mon[53831]: pgmap v348: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:29.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:29 vm01 ceph-mon[52793]: pgmap v348: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:31.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:31 vm04 ceph-mon[53158]: pgmap v349: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:31.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:31 vm05 ceph-mon[53831]: pgmap v349: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:31.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:31 vm01 ceph-mon[52793]: pgmap v349: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:33.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:33 vm04 ceph-mon[53158]: pgmap v350: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:33.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:33 vm05 ceph-mon[53831]: pgmap v350: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:33.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:33 vm01 ceph-mon[52793]: pgmap v350: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:35.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:51:34 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:51:35.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:35 vm01 ceph-mon[52793]: pgmap v351: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:35 vm04 ceph-mon[53158]: pgmap v351: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:35.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:35 vm05 ceph-mon[53831]: pgmap v351: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:36 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:36.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:36 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:36.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:51:36 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:51:36.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:36 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:37.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:37 vm01 ceph-mon[52793]: pgmap v352: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:37.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:37 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:37.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:37 vm04 ceph-mon[53158]: pgmap v352: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:37.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:37 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:37.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:37 vm05 ceph-mon[53831]: pgmap v352: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:37.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:37 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:39.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:39 vm01 ceph-mon[52793]: pgmap v353: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:39.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:39 vm04 ceph-mon[53158]: pgmap v353: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:39.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:39 vm05 ceph-mon[53831]: pgmap v353: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:40.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:51:40.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:40.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:40.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:40.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:40.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:51:40.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:51:40.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:40.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:51:40.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:40.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:40.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:40.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:40.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:51:40.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:51:40.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:40.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:51:40.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:40.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:40.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:40.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:40.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:51:40.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:51:40.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:51:41.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:41 vm01 ceph-mon[52793]: pgmap v354: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:41.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:41 vm04 ceph-mon[53158]: pgmap v354: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:41.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:41 vm05 ceph-mon[53831]: pgmap v354: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:43.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:43 vm01 ceph-mon[52793]: pgmap v355: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:43.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:43 vm04 ceph-mon[53158]: pgmap v355: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:43.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:43 vm05 ceph-mon[53831]: pgmap v355: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:45.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:51:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:51:45.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:45 vm01 ceph-mon[52793]: pgmap v356: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:45.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:45 vm04 ceph-mon[53158]: pgmap v356: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:45.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:45 vm05 ceph-mon[53831]: pgmap v356: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:46.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:46 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:46.570 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:51:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:51:46.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:46 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:46.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:46 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:47.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:47 vm01 ceph-mon[52793]: pgmap v357: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:47.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:47 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:47.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:47 vm04 ceph-mon[53158]: pgmap v357: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:47.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:47 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:47.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:47 vm05 ceph-mon[53831]: pgmap v357: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:47.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:47 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:48.727 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:48 vm01 ceph-mon[52793]: pgmap v358: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:48.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:48 vm04 ceph-mon[53158]: pgmap v358: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:48.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:48 vm05 ceph-mon[53831]: pgmap v358: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:50.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:50 vm04 ceph-mon[53158]: pgmap v359: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:50.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:50 vm05 ceph-mon[53831]: pgmap v359: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:50.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:50 vm01 ceph-mon[52793]: pgmap v359: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:52.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:52 vm04 ceph-mon[53158]: pgmap v360: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:52.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:52 vm05 ceph-mon[53831]: pgmap v360: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:52.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:52 vm01 ceph-mon[52793]: pgmap v360: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:54.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:54 vm04 ceph-mon[53158]: pgmap v361: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:54.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:54 vm05 ceph-mon[53831]: pgmap v361: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:54.963 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:54 vm01 ceph-mon[52793]: pgmap v361: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:55.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:51:54 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:51:55.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:55 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:55.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:55 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:55.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:55 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:56.534 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:51:56 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:51:56.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:56 vm04 ceph-mon[53158]: pgmap v362: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:56.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:56 vm05 ceph-mon[53831]: pgmap v362: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:56.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:56 vm01 ceph-mon[52793]: pgmap v362: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:51:57.791 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:57 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:57.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:57 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:57.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:57 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:51:58.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:51:58 vm05 ceph-mon[53831]: pgmap v363: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:58.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:51:58 vm01 ceph-mon[52793]: pgmap v363: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:51:59.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:51:58 vm04 ceph-mon[53158]: pgmap v363: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:00.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:00 vm05 ceph-mon[53831]: pgmap v364: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:00.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:00 vm01 ceph-mon[52793]: pgmap v364: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:01.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:00 vm04 ceph-mon[53158]: pgmap v364: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:02.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:02 vm05 ceph-mon[53831]: pgmap v365: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:02.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:02 vm01 ceph-mon[52793]: pgmap v365: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:03.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:02 vm04 ceph-mon[53158]: pgmap v365: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:04.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:04 vm05 ceph-mon[53831]: pgmap v366: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:04.966 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:04 vm01 ceph-mon[52793]: pgmap v366: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:05.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:04 vm04 ceph-mon[53158]: pgmap v366: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:05.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:52:04 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:52:05.820 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:05 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:05.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:05 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:06.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:05 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:06.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:52:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:52:06.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:06 vm01 ceph-mon[52793]: pgmap v367: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:07.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:06 vm04 ceph-mon[53158]: pgmap v367: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:07.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:06 vm05 ceph-mon[53831]: pgmap v367: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:07.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:07 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:08.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:07 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:08.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:07 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:08.977 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:08 vm01 ceph-mon[52793]: pgmap v368: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:09.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:08 vm04 ceph-mon[53158]: pgmap v368: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:09.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:08 vm05 ceph-mon[53831]: pgmap v368: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:11.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:10 vm04 ceph-mon[53158]: pgmap v369: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:11.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:10 vm05 ceph-mon[53831]: pgmap v369: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:11.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:10 vm01 ceph-mon[52793]: pgmap v369: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:13.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:12 vm04 ceph-mon[53158]: pgmap v370: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:13.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:12 vm05 ceph-mon[53831]: pgmap v370: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:13.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:12 vm01 ceph-mon[52793]: pgmap v370: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:15.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:14 vm04 ceph-mon[53158]: pgmap v371: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:15.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:14 vm05 ceph-mon[53831]: pgmap v371: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:15.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:52:14 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:52:15.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:14 vm01 ceph-mon[52793]: pgmap v371: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:16.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:15 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:16.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:15 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:16.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:15 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:16.570 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:52:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:52:17.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:16 vm04 ceph-mon[53158]: pgmap v372: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:17.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:16 vm05 ceph-mon[53831]: pgmap v372: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:17.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:16 vm01 ceph-mon[52793]: pgmap v372: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:18.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:17 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:18.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:17 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:18.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:17 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:19.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:18 vm04 ceph-mon[53158]: pgmap v373: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:19.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:18 vm05 ceph-mon[53831]: pgmap v373: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:19.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:18 vm01 ceph-mon[52793]: pgmap v373: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:21.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:20 vm04 ceph-mon[53158]: pgmap v374: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:21.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:20 vm05 ceph-mon[53831]: pgmap v374: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:21.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:20 vm01 ceph-mon[52793]: pgmap v374: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:23.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:22 vm04 ceph-mon[53158]: pgmap v375: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:23.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:22 vm05 ceph-mon[53831]: pgmap v375: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:23.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:22 vm01 ceph-mon[52793]: pgmap v375: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:25.030 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:52:24 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:52:25.030 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:24 vm01 ceph-mon[52793]: pgmap v376: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:25.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:24 vm04 ceph-mon[53158]: pgmap v376: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:25.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:24 vm05 ceph-mon[53831]: pgmap v376: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:26.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:25 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:26.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:25 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:26.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:25 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:26.570 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:52:26 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:52:27.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:26 vm04 ceph-mon[53158]: pgmap v377: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:27.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:26 vm05 ceph-mon[53831]: pgmap v377: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:27.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:26 vm01 ceph-mon[52793]: pgmap v377: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:28.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:27 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:28.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:27 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:28.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:27 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:29.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:28 vm04 ceph-mon[53158]: pgmap v378: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:29.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:28 vm05 ceph-mon[53831]: pgmap v378: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:29.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:28 vm01 ceph-mon[52793]: pgmap v378: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:31.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:30 vm04 ceph-mon[53158]: pgmap v379: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:31.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:30 vm05 ceph-mon[53831]: pgmap v379: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:31.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:30 vm01 ceph-mon[52793]: pgmap v379: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:33.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:32 vm05 ceph-mon[53831]: pgmap v380: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:33.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:32 vm01 ceph-mon[52793]: pgmap v380: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:33.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:32 vm04 ceph-mon[53158]: pgmap v380: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:35.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:34 vm05 ceph-mon[53831]: pgmap v381: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:35.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:52:34 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:52:35.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:34 vm01 ceph-mon[52793]: pgmap v381: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:35.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:34 vm04 ceph-mon[53158]: pgmap v381: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:36.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:35 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:36.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:35 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:36.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:35 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:36.570 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:52:36 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:52:37.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:36 vm05 ceph-mon[53831]: pgmap v382: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:37.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:36 vm01 ceph-mon[52793]: pgmap v382: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:37.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:36 vm04 ceph-mon[53158]: pgmap v382: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:38.070 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:37 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:38.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:37 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:38.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:37 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:39.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:38 vm01 ceph-mon[52793]: pgmap v383: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:39.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:38 vm04 ceph-mon[53158]: pgmap v383: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:39.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:38 vm05 ceph-mon[53831]: pgmap v383: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:41.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:40 vm01 ceph-mon[52793]: pgmap v384: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:41.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:52:41.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:52:41.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:52:41.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:52:41.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:52:41.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:52:41.228 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:40 vm01 ceph-mon[52793]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:52:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:40 vm04 ceph-mon[53158]: pgmap v384: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:52:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:52:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:52:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:52:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:52:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:52:41.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:40 vm04 ceph-mon[53158]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:52:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:40 vm05 ceph-mon[53831]: pgmap v384: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:52:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:52:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:52:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:52:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:52:41.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:52:41.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:40 vm05 ceph-mon[53831]: from='mgr.14150 192.168.123.101:0/203938607' entity='mgr.x' 2026-03-09T17:52:43.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:43 vm04 ceph-mon[53158]: pgmap v385: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:43.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:43 vm05 ceph-mon[53831]: pgmap v385: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:43.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:43 vm01 ceph-mon[52793]: pgmap v385: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:45.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:45 vm04 ceph-mon[53158]: pgmap v386: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:45.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:45 vm05 ceph-mon[53831]: pgmap v386: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:45.477 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:52:44 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:52:45.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:45 vm01 ceph-mon[52793]: pgmap v386: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:46.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:46 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:46.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:46 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:46.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:52:46 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:52:46.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:46 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:47.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:47 vm04 ceph-mon[53158]: pgmap v387: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:47.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:47 vm05 ceph-mon[53831]: pgmap v387: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:47.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:47 vm01 ceph-mon[52793]: pgmap v387: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:48.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:48 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:48.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:48 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:48.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:48 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:49.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:49 vm04 ceph-mon[53158]: pgmap v388: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:49.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:49 vm05 ceph-mon[53831]: pgmap v388: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:49.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:49 vm01 ceph-mon[52793]: pgmap v388: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:51.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:51 vm04 ceph-mon[53158]: pgmap v389: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:51.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:51 vm05 ceph-mon[53831]: pgmap v389: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:51.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:51 vm01 ceph-mon[52793]: pgmap v389: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:53.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:53 vm04 ceph-mon[53158]: pgmap v390: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:53.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:53 vm05 ceph-mon[53831]: pgmap v390: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:53.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:53 vm01 ceph-mon[52793]: pgmap v390: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:55.279 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:52:55 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:52:55.280 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:55 vm01 ceph-mon[52793]: pgmap v391: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:55.291 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:55 vm04 ceph-mon[53158]: pgmap v391: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:55.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:55 vm05 ceph-mon[53831]: pgmap v391: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:56.321 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:56 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:56.321 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:52:56 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:52:56.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:56 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:56.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:56 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:57.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:57 vm05 ceph-mon[53831]: pgmap v392: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:57.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:57 vm01 ceph-mon[52793]: pgmap v392: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:57.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:57 vm04 ceph-mon[53158]: pgmap v392: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:52:58.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:58 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:58.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:58 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:58.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:58 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:52:59.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:52:59 vm05 ceph-mon[53831]: pgmap v393: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:59.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:52:59 vm01 ceph-mon[52793]: pgmap v393: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:52:59.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:52:59 vm04 ceph-mon[53158]: pgmap v393: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:53:01.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:01 vm05 ceph-mon[53831]: pgmap v394: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:01.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:01 vm01 ceph-mon[52793]: pgmap v394: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:01.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:01 vm04 ceph-mon[53158]: pgmap v394: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:03.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:03 vm05 ceph-mon[53831]: pgmap v395: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:53:03.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:03 vm01 ceph-mon[52793]: pgmap v395: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:53:03.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:03 vm04 ceph-mon[53158]: pgmap v395: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:53:05.320 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:05 vm05 ceph-mon[53831]: pgmap v396: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:05.477 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:05 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:53:05.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:05 vm01 ceph-mon[52793]: pgmap v396: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:05.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:05 vm04 ceph-mon[53158]: pgmap v396: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:06.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:06 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:06.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:06 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:06.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:06 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:06.570 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:06 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:53:07.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:07 vm01 ceph-mon[52793]: pgmap v397: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:07.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:07 vm04 ceph-mon[53158]: pgmap v397: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:07.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:07 vm05 ceph-mon[53831]: pgmap v397: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:08.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:08 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:08.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:08 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:08.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:08 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:09.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:09 vm01 ceph-mon[52793]: pgmap v398: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:53:09.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:09 vm04 ceph-mon[53158]: pgmap v398: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:53:09.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:09 vm05 ceph-mon[53831]: pgmap v398: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:53:11.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:11 vm01 ceph-mon[52793]: pgmap v399: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:11.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:11 vm04 ceph-mon[53158]: pgmap v399: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:11.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:11 vm05 ceph-mon[53831]: pgmap v399: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:13.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:13 vm01 ceph-mon[52793]: pgmap v400: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:53:13.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:13 vm04 ceph-mon[53158]: pgmap v400: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:53:13.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:13 vm05 ceph-mon[53831]: pgmap v400: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:53:15.477 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:15 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug there is no tcmu-runner data available 2026-03-09T17:53:15.478 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:15 vm01 ceph-mon[52793]: pgmap v401: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:15.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:15 vm04 ceph-mon[53158]: pgmap v401: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:15.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:15 vm05 ceph-mon[53831]: pgmap v401: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr:Note: switching to '569c3e99c9b32a51b4eaf08731c728f4513ed589'. 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr:state without impacting any branches by switching back to a branch. 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr: git switch -c 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr:Or undo this operation with: 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr: git switch - 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:53:15.857 INFO:teuthology.orchestra.run.vm05.stderr:HEAD is now at 569c3e99c9b qa/rgw: bucket notifications use pynose 2026-03-09T17:53:15.862 DEBUG:teuthology.orchestra.run.vm05:> cp -- /home/ubuntu/cephtest/clone.client.2/src/test/cli-integration/rbd/rest_api_delete.t /home/ubuntu/cephtest/archive/cram.client.2 2026-03-09T17:53:15.919 INFO:tasks.cram:Running tests for client.0... 2026-03-09T17:53:15.919 DEBUG:teuthology.orchestra.run.vm01:> CEPH_REF=master CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.0/*.t 2026-03-09T17:53:16.213 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:16.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/642714048' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:53:16.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/642714048' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:53:16.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/642714048' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:53:16.214 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/642714048' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/2403195361"}]: dispatch 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:15 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:53:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:15 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:53:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (LUN.allocate) created datapool/block1 successfully 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block1' to LIO backstore user:rbd 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (LUN.add_dev_to_lio) Successfully added datapool/block1 to LIO 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (LUN.allocate) added 'datapool/block1' to LIO and config object 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:53:16] "PUT /api/_disk/datapool/block1 HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:53:16] "PUT /api/_disk/datapool/block1 HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _disk update on localhost, successful 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:16] "PUT /api/disk/datapool/block1 HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:16] "PUT /api/disk/datapool/block1 HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Gateway.create_target) created an iscsi target with iqn of 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:16] "PUT /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:16] "PUT /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug gateway validation needed for vm01.local 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug An additional 1 tpg's are required 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Gateway.create_tpg) created TPG '1' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug iqn.2003-01.com.redhat.iscsi-gw:ceph-gw - Adding the IP to the enabled tpg, allowing iSCSI logins 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:53:16] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:53:16] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _gateway update on localhost, successful 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:16] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-09T17:53:16.214 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:16] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug gateway validation needed for vm05.local 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _gateway update on vm05.local, successful 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug An additional 1 tpg's are required 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Gateway.create_tpg) created TPG '2' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _gateway update on vm01.local, successful 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:16] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:16] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block1' to LIO backstore user:rbd 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (LUN.add_dev_to_lio) Successfully added datapool/block1 to LIO 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Change detected - internal 17 / xattr 18 refreshing 2026-03-09T17:53:16.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Setup group ano2 for datapool.block1 on tpg 2 (state 1, owner False, failover type 1) 2026-03-09T17:53:16.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:16 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:16.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:16 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/642714048' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:53:16.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:16 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/642714048' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:53:16.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:16 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/642714048' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:53:16.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:16 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/642714048' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/2403195361"}]: dispatch 2026-03-09T17:53:16.541 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:16 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:16.541 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:16 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/642714048' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:53:16.541 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:16 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/642714048' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:53:16.541 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:16 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/642714048' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:53:16.541 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:16 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/642714048' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/2403195361"}]: dispatch 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Gateway.create_tpg) created TPG '1' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Gateway.create_tpg) created TPG '2' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Gateway.create_target) created an iscsi target with iqn of 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug iqn.2003-01.com.redhat.iscsi-gw:ceph-gw - Adding the IP to the enabled tpg, allowing iSCSI logins 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug there is no tcmu-runner data available 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.541 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Setup group ao for datapool.block1 on tpg 1 (state 0, owner True, failover type 1) 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _targetlun update on vm01.local, successful 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _targetlun update on vm05.local, successful 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:16] "PUT /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:16] "PUT /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:53:16] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:53:16] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Client.define_client) iqn.1994-05.com.redhat:client added successfully 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:53:16] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:53:16] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _client update on localhost, successful 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _client update on vm05.local, successful 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:16] "PUT /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:16] "PUT /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:53:16] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:53:16] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Client.add_lun) added image 'datapool/block1' to iqn.1994-05.com.redhat:client 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::1 - - [09/Mar/2026 17:53:16] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::1 - - [09/Mar/2026 17:53:16] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:16.728 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _clientlun update on localhost, successful 2026-03-09T17:53:16.736 INFO:tasks.cram.client.0.vm01.stdout:/home/ubuntu/cephtest/archive/cram.client.0/rest_api_create.t: passed 2026-03-09T17:53:16.736 INFO:tasks.cram.client.0.vm01.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-09T17:53:16.740 INFO:tasks.cram:Running tests for client.1... 2026-03-09T17:53:16.740 DEBUG:teuthology.orchestra.run.vm04:> CEPH_REF=master CEPH_ID="1" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.1/*.t 2026-03-09T17:53:16.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block1' to LIO backstore user:rbd 2026-03-09T17:53:16.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (LUN.add_dev_to_lio) Successfully added datapool/block1 to LIO 2026-03-09T17:53:16.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (LUN.allocate) added block1 to LIO using wwn '839a137d-c48e-420c-9600-3f9670ddc9a6' defined by vm01.local 2026-03-09T17:53:16.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-09T17:53:16.820 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug Setup group ano2 for datapool.block1 on tpg 2 (state 1, owner False, failover type 1) 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug Setup group ao for datapool.block1 on tpg 1 (state 0, owner True, failover type 1) 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Client.define_client) iqn.1994-05.com.redhat:client added successfully 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Client.add_lun) added image 'datapool/block1' to iqn.1994-05.com.redhat:client 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:16.821 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:16 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.101 - - [09/Mar/2026 17:53:16] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:17.140 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug _clientlun update on vm05.local, successful 2026-03-09T17:53:17.140 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:16] "PUT /api/clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:17.140 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:16 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:16] "PUT /api/clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:17.403 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:17 vm04 ceph-mon[53158]: pgmap v402: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:17.403 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:17 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/311908651' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:53:17.404 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:17 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/311908651' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:53:17.404 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:17 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/311908651' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:53:17.404 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:17 vm04 ceph-mon[53158]: from='client.? 192.168.123.101:0/311908651' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/642714048"}]: dispatch 2026-03-09T17:53:17.404 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:17 vm04 ceph-mon[53158]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/642714048"}]: dispatch 2026-03-09T17:53:17.404 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:17 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3136121186' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:53:17.404 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:17 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3136121186' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:53:17.404 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:17 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3136121186' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:53:17.404 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:17 vm04 ceph-mon[53158]: from='client.? 192.168.123.105:0/3136121186' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.105:0/3205701074"}]: dispatch 2026-03-09T17:53:17.404 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:17 vm04 ceph-mon[53158]: from='client.? ' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.105:0/3205701074"}]: dispatch 2026-03-09T17:53:17.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:17 vm01 ceph-mon[52793]: pgmap v402: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:17.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:17 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/311908651' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:53:17.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:17 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/311908651' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:53:17.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:17 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/311908651' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:53:17.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:17 vm01 ceph-mon[52793]: from='client.? 192.168.123.101:0/311908651' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/642714048"}]: dispatch 2026-03-09T17:53:17.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:17 vm01 ceph-mon[52793]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/642714048"}]: dispatch 2026-03-09T17:53:17.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:17 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3136121186' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:53:17.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:17 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3136121186' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:53:17.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:17 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3136121186' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:53:17.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:17 vm01 ceph-mon[52793]: from='client.? 192.168.123.105:0/3136121186' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.105:0/3205701074"}]: dispatch 2026-03-09T17:53:17.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:17 vm01 ceph-mon[52793]: from='client.? ' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.105:0/3205701074"}]: dispatch 2026-03-09T17:53:17.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:17 vm05 ceph-mon[53831]: pgmap v402: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:17.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:17 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/311908651' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:53:17.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:17 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/311908651' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:53:17.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:17 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/311908651' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:53:17.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:17 vm05 ceph-mon[53831]: from='client.? 192.168.123.101:0/311908651' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/642714048"}]: dispatch 2026-03-09T17:53:17.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:17 vm05 ceph-mon[53831]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/642714048"}]: dispatch 2026-03-09T17:53:17.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:17 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3136121186' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-09T17:53:17.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:17 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3136121186' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-09T17:53:17.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:17 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3136121186' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-09T17:53:17.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:17 vm05 ceph-mon[53831]: from='client.? 192.168.123.105:0/3136121186' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.105:0/3205701074"}]: dispatch 2026-03-09T17:53:17.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:17 vm05 ceph-mon[53831]: from='client.? ' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.105:0/3205701074"}]: dispatch 2026-03-09T17:53:18.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:18 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:18.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:18 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:18.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:18 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:19.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:19 vm01 ceph-mon[52793]: pgmap v403: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:53:19.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:19 vm04 ceph-mon[53158]: pgmap v403: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:53:19.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:19 vm05 ceph-mon[53831]: pgmap v403: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-09T17:53:21.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:21 vm01 ceph-mon[52793]: pgmap v404: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:21.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:21 vm04 ceph-mon[53158]: pgmap v404: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:21.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:21 vm05 ceph-mon[53831]: pgmap v404: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-09T17:53:23.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:23 vm01 ceph-mon[52793]: pgmap v405: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:23.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:23 vm04 ceph-mon[53158]: pgmap v405: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:23.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:23 vm05 ceph-mon[53831]: pgmap v405: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:25.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:25 vm01 ceph-mon[52793]: pgmap v406: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:25.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:25 vm04 ceph-mon[53158]: pgmap v406: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:25.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:25 vm05 ceph-mon[53831]: pgmap v406: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:26.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:26 vm01 ceph-mon[52793]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:26.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:26 vm04 ceph-mon[53158]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:26.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:26 vm05 ceph-mon[53831]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:27.149 INFO:tasks.cram.client.1.vm04.stdout:/home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t: passed 2026-03-09T17:53:27.149 INFO:tasks.cram.client.1.vm04.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-09T17:53:27.154 INFO:tasks.cram:Running tests for client.2... 2026-03-09T17:53:27.154 DEBUG:teuthology.orchestra.run.vm05:> CEPH_REF=master CEPH_ID="2" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.2/*.t 2026-03-09T17:53:27.430 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:27 vm05 ceph-mon[53831]: pgmap v407: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:53:27] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:53:27] "GET /api/config HTTP/1.1" 200 - 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:53:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:53:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug _client update on vm01.local, successful 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Client.delete) deleted NodeACL for iqn.1994-05.com.redhat:client 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::1 - - [09/Mar/2026 17:53:27] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::1 - - [09/Mar/2026 17:53:27] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug _client update on localhost, successful 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:27] "DELETE /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:27] "DELETE /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug _targetlun update on vm01.local, successful 2026-03-09T17:53:27.432 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug LUN unmap request received, config commit to be performed by vm05.local 2026-03-09T17:53:27.468 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:27.468 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:27.468 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Client.delete) deleted NodeACL for iqn.1994-05.com.redhat:client 2026-03-09T17:53:27.468 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:27.468 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-09T17:53:27.468 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:27.468 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:27.468 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug LUN unmap request received, config commit to be performed by vm05.local 2026-03-09T17:53:27.468 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:27.468 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:27.468 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-mon[52793]: pgmap v407: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:27.541 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:27 vm04 ceph-mon[53158]: pgmap v407: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:27.691 INFO:tasks.cram.client.2.vm05.stdout:/home/ubuntu/cephtest/archive/cram.client.2/rest_api_delete.t: passed 2026-03-09T17:53:27.691 INFO:tasks.cram.client.2.vm05.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-09T17:53:27.693 DEBUG:teuthology.orchestra.run.vm01:> test -f /home/ubuntu/cephtest/archive/cram.client.0/rest_api_create.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.0/rest_api_create.t 2026-03-09T17:53:27.710 DEBUG:teuthology.orchestra.run.vm01:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.0 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.0 2026-03-09T17:53:27.727 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Change detected - internal 22 / xattr 23 refreshing 2026-03-09T17:53:27.727 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:27.727 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:27.727 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-09T17:53:27.727 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Removing target configuration 2026-03-09T17:53:27.727 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:27.727 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:27 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug _targetlun update on vm05.local, successful 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:27] "DELETE /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:27] "DELETE /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug _target update on vm01.local, successful 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug Removing target configuration 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug _target update on vm05.local, successful 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:27] "DELETE /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:27] "DELETE /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug LUN deletion request received, rbd removal to be performed by vm05.local 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "DELETE /api/_disk/datapool/block1 HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:192.168.123.105 - - [09/Mar/2026 17:53:27] "DELETE /api/_disk/datapool/block1 HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug _disk update on vm05.local, successful 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:27] "DELETE /api/disk/datapool/block1 HTTP/1.1" 200 - 2026-03-09T17:53:27.822 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:27 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: ::ffff:127.0.0.1 - - [09/Mar/2026 17:53:27] "DELETE /api/disk/datapool/block1 HTTP/1.1" 200 - 2026-03-09T17:53:28.093 DEBUG:teuthology.orchestra.run.vm04:> test -f /home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t 2026-03-09T17:53:28.110 DEBUG:teuthology.orchestra.run.vm04:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.1 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.1 2026-03-09T17:53:28.212 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:28 vm04 ceph-mon[53158]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:28.471 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:28 vm01 ceph-mon[52793]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:28.549 DEBUG:teuthology.orchestra.run.vm05:> test -f /home/ubuntu/cephtest/archive/cram.client.2/rest_api_delete.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.2/rest_api_delete.t 2026-03-09T17:53:28.565 DEBUG:teuthology.orchestra.run.vm05:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.2 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.2 2026-03-09T17:53:28.570 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:28 vm05 ceph-mon[53831]: from='client.24377 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-09T17:53:28.727 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:28 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Change detected - internal 24 / xattr 26 refreshing 2026-03-09T17:53:28.886 DEBUG:teuthology.run_tasks:Unwinding manager ceph_iscsi_client 2026-03-09T17:53:28.888 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-09T17:53:28.890 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-09T17:53:28.890 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T17:53:28.892 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T17:53:28.893 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T17:53:28.929 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T17:53:28.929 DEBUG:teuthology.orchestra.run.vm01:> 2026-03-09T17:53:28.929 DEBUG:teuthology.orchestra.run.vm01:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T17:53:28.929 DEBUG:teuthology.orchestra.run.vm01:> sudo yum -y remove $d || true 2026-03-09T17:53:28.929 DEBUG:teuthology.orchestra.run.vm01:> done 2026-03-09T17:53:28.934 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T17:53:28.934 DEBUG:teuthology.orchestra.run.vm04:> 2026-03-09T17:53:28.934 DEBUG:teuthology.orchestra.run.vm04:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T17:53:28.934 DEBUG:teuthology.orchestra.run.vm04:> sudo yum -y remove $d || true 2026-03-09T17:53:28.934 DEBUG:teuthology.orchestra.run.vm04:> done 2026-03-09T17:53:28.939 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T17:53:28.939 DEBUG:teuthology.orchestra.run.vm05:> 2026-03-09T17:53:28.939 DEBUG:teuthology.orchestra.run.vm05:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T17:53:28.939 DEBUG:teuthology.orchestra.run.vm05:> sudo yum -y remove $d || true 2026-03-09T17:53:28.939 DEBUG:teuthology.orchestra.run.vm05:> done 2026-03-09T17:53:29.135 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:29.136 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:29.136 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T17:53:29.136 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:29.136 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T17:53:29.136 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T17:53:29.136 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-09T17:53:29.136 INFO:teuthology.orchestra.run.vm01.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T17:53:29.136 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:29.136 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T17:53:29.136 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:29.136 INFO:teuthology.orchestra.run.vm01.stdout:Remove 2 Packages 2026-03-09T17:53:29.136 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:29.136 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 39 M 2026-03-09T17:53:29.136 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T17:53:29.138 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T17:53:29.138 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T17:53:29.142 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:29.142 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:29.142 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-09T17:53:29.142 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:29.142 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-09T17:53:29.142 INFO:teuthology.orchestra.run.vm05.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T17:53:29.142 INFO:teuthology.orchestra.run.vm05.stdout:Removing unused dependencies: 2026-03-09T17:53:29.142 INFO:teuthology.orchestra.run.vm05.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T17:53:29.142 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:29.142 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-09T17:53:29.142 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:29.142 INFO:teuthology.orchestra.run.vm05.stdout:Remove 2 Packages 2026-03-09T17:53:29.142 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:29.143 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 39 M 2026-03-09T17:53:29.143 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-09T17:53:29.145 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-09T17:53:29.145 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-09T17:53:29.152 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T17:53:29.152 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T17:53:29.154 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:29.154 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:29.155 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T17:53:29.155 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:29.155 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T17:53:29.155 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T17:53:29.155 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T17:53:29.155 INFO:teuthology.orchestra.run.vm04.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T17:53:29.155 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:29.155 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T17:53:29.155 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:29.155 INFO:teuthology.orchestra.run.vm04.stdout:Remove 2 Packages 2026-03-09T17:53:29.155 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:29.155 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 39 M 2026-03-09T17:53:29.155 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T17:53:29.157 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T17:53:29.157 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T17:53:29.158 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-09T17:53:29.158 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-09T17:53:29.172 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T17:53:29.172 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T17:53:29.183 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T17:53:29.207 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:29.207 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:29.207 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T17:53:29.207 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T17:53:29.207 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T17:53:29.207 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:29.210 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:29.215 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:29 vm05 ceph-mon[53831]: pgmap v408: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:29.215 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:29 vm04 ceph-mon[53158]: pgmap v408: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:29.216 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-09T17:53:29.220 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:29.234 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T17:53:29.241 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:29.241 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:29.241 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T17:53:29.241 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T17:53:29.241 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T17:53:29.241 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:29.241 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T17:53:29.243 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:29.252 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:29.267 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T17:53:29.269 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:29.269 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:29.269 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T17:53:29.269 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T17:53:29.269 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T17:53:29.269 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:29.269 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:29.278 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:29.294 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T17:53:29.305 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T17:53:29.305 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:29.343 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T17:53:29.343 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:29.351 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T17:53:29.351 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:29.351 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T17:53:29.351 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T17:53:29.351 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:29.351 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:29.387 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T17:53:29.387 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:29.391 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T17:53:29.391 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:29.392 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-09T17:53:29.392 INFO:teuthology.orchestra.run.vm05.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T17:53:29.392 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:29.392 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:29.441 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T17:53:29.441 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:29.441 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T17:53:29.441 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T17:53:29.441 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:29.441 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:29.477 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:29 vm01 ceph-mon[52793]: pgmap v408: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:29.556 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:29.556 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:29.556 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T17:53:29.557 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:29.557 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T17:53:29.557 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T17:53:29.557 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-09T17:53:29.557 INFO:teuthology.orchestra.run.vm01.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T17:53:29.557 INFO:teuthology.orchestra.run.vm01.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T17:53:29.557 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T17:53:29.557 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:29.557 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T17:53:29.557 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:29.557 INFO:teuthology.orchestra.run.vm01.stdout:Remove 4 Packages 2026-03-09T17:53:29.557 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:29.557 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 212 M 2026-03-09T17:53:29.557 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T17:53:29.560 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T17:53:29.560 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T17:53:29.586 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T17:53:29.587 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T17:53:29.592 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout:Removing unused dependencies: 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout:Remove 4 Packages 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 212 M 2026-03-09T17:53:29.593 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-09T17:53:29.596 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-09T17:53:29.596 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-09T17:53:29.620 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-09T17:53:29.621 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-09T17:53:29.648 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout:Remove 4 Packages 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 212 M 2026-03-09T17:53:29.649 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T17:53:29.650 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T17:53:29.652 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T17:53:29.652 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T17:53:29.656 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T17:53:29.658 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T17:53:29.661 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T17:53:29.677 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T17:53:29.678 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T17:53:29.678 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T17:53:29.685 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-09T17:53:29.691 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T17:53:29.693 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T17:53:29.696 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T17:53:29.712 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T17:53:29.744 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T17:53:29.749 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T17:53:29.749 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T17:53:29.749 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T17:53:29.749 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T17:53:29.751 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T17:53:29.753 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T17:53:29.757 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T17:53:29.774 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T17:53:29.784 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T17:53:29.784 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T17:53:29.784 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T17:53:29.784 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T17:53:29.799 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T17:53:29.800 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:29.800 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T17:53:29.800 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T17:53:29.800 INFO:teuthology.orchestra.run.vm01.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T17:53:29.800 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:29.800 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:29.832 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T17:53:29.832 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:29.832 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-09T17:53:29.832 INFO:teuthology.orchestra.run.vm05.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T17:53:29.832 INFO:teuthology.orchestra.run.vm05.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T17:53:29.832 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:29.832 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:29.846 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T17:53:29.846 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T17:53:29.846 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T17:53:29.846 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T17:53:29.895 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T17:53:29.895 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:29.895 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T17:53:29.895 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T17:53:29.895 INFO:teuthology.orchestra.run.vm04.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T17:53:29.896 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:29.896 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:30.018 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout:Remove 8 Packages 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 28 M 2026-03-09T17:53:30.019 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T17:53:30.022 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T17:53:30.022 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T17:53:30.044 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout:Removing unused dependencies: 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout:Remove 8 Packages 2026-03-09T17:53:30.045 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:30.046 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 28 M 2026-03-09T17:53:30.046 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-09T17:53:30.048 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T17:53:30.048 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T17:53:30.048 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-09T17:53:30.048 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-09T17:53:30.071 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-09T17:53:30.071 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-09T17:53:30.092 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T17:53:30.097 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T17:53:30.101 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T17:53:30.102 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T17:53:30.103 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout:Remove 8 Packages 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 28 M 2026-03-09T17:53:30.104 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T17:53:30.105 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T17:53:30.107 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T17:53:30.107 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T17:53:30.108 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T17:53:30.110 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T17:53:30.111 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-09T17:53:30.116 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T17:53:30.119 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T17:53:30.121 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T17:53:30.124 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T17:53:30.127 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T17:53:30.128 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T17:53:30.132 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T17:53:30.133 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T17:53:30.134 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T17:53:30.134 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:30.134 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T17:53:30.134 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T17:53:30.134 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T17:53:30.134 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:30.134 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T17:53:30.140 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T17:53:30.150 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T17:53:30.150 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:30.150 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T17:53:30.151 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T17:53:30.151 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T17:53:30.151 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:30.151 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T17:53:30.159 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T17:53:30.163 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T17:53:30.163 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:30.163 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T17:53:30.163 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T17:53:30.163 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T17:53:30.163 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:30.164 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T17:53:30.176 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T17:53:30.181 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T17:53:30.183 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T17:53:30.183 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:30.183 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T17:53:30.183 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T17:53:30.183 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T17:53:30.183 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:30.184 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T17:53:30.185 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T17:53:30.187 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T17:53:30.190 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T17:53:30.192 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T17:53:30.195 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T17:53:30.219 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T17:53:30.220 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:30.220 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T17:53:30.220 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T17:53:30.220 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T17:53:30.220 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:30.220 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T17:53:30.228 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T17:53:30.254 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T17:53:30.254 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:30.254 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T17:53:30.255 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T17:53:30.255 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T17:53:30.255 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:30.255 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T17:53:30.265 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T17:53:30.266 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T17:53:30.266 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T17:53:30.266 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T17:53:30.266 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T17:53:30.266 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T17:53:30.266 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T17:53:30.266 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T17:53:30.274 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T17:53:30.274 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T17:53:30.274 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T17:53:30.274 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T17:53:30.274 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T17:53:30.274 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T17:53:30.274 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T17:53:30.274 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T17:53:30.322 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T17:53:30.322 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:30.322 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T17:53:30.322 INFO:teuthology.orchestra.run.vm01.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:30.322 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:30.322 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:30.322 INFO:teuthology.orchestra.run.vm01.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T17:53:30.322 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T17:53:30.322 INFO:teuthology.orchestra.run.vm01.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T17:53:30.323 INFO:teuthology.orchestra.run.vm01.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T17:53:30.323 INFO:teuthology.orchestra.run.vm01.stdout: zip-3.0-35.el9.x86_64 2026-03-09T17:53:30.323 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:30.323 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:30.327 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T17:53:30.327 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:30.327 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-09T17:53:30.327 INFO:teuthology.orchestra.run.vm05.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:30.327 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:30.327 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:30.327 INFO:teuthology.orchestra.run.vm05.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T17:53:30.327 INFO:teuthology.orchestra.run.vm05.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T17:53:30.327 INFO:teuthology.orchestra.run.vm05.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T17:53:30.327 INFO:teuthology.orchestra.run.vm05.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T17:53:30.327 INFO:teuthology.orchestra.run.vm05.stdout: zip-3.0-35.el9.x86_64 2026-03-09T17:53:30.327 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:30.327 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:30.359 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T17:53:30.359 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T17:53:30.359 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T17:53:30.359 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T17:53:30.359 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T17:53:30.359 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T17:53:30.359 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T17:53:30.359 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T17:53:30.401 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:30 vm01 ceph-mon[52793]: pgmap v409: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:30.415 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T17:53:30.415 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:30.415 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T17:53:30.415 INFO:teuthology.orchestra.run.vm04.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:30.415 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:30.415 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:30.415 INFO:teuthology.orchestra.run.vm04.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T17:53:30.415 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T17:53:30.415 INFO:teuthology.orchestra.run.vm04.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T17:53:30.415 INFO:teuthology.orchestra.run.vm04.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T17:53:30.415 INFO:teuthology.orchestra.run.vm04.stdout: zip-3.0-35.el9.x86_64 2026-03-09T17:53:30.415 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:30.415 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:30.417 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:30 vm05 ceph-mon[53831]: pgmap v409: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:30.520 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:30 vm04 ceph-mon[53158]: pgmap v409: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-09T17:53:30.535 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:30.540 INFO:teuthology.orchestra.run.vm01.stdout:=========================================================================================== 2026-03-09T17:53:30.540 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T17:53:30.540 INFO:teuthology.orchestra.run.vm01.stdout:=========================================================================================== 2026-03-09T17:53:30.540 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T17:53:30.540 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T17:53:30.540 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-09T17:53:30.540 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T17:53:30.540 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T17:53:30.540 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T17:53:30.540 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T17:53:30.540 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T17:53:30.541 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout:=========================================================================================== 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout:Remove 102 Packages 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 613 M 2026-03-09T17:53:30.542 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T17:53:30.548 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:30.553 INFO:teuthology.orchestra.run.vm05.stdout:=========================================================================================== 2026-03-09T17:53:30.553 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-09T17:53:30.553 INFO:teuthology.orchestra.run.vm05.stdout:=========================================================================================== 2026-03-09T17:53:30.553 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-09T17:53:30.553 INFO:teuthology.orchestra.run.vm05.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T17:53:30.553 INFO:teuthology.orchestra.run.vm05.stdout:Removing dependent packages: 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout:Removing unused dependencies: 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T17:53:30.554 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout:=========================================================================================== 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout:Remove 102 Packages 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 613 M 2026-03-09T17:53:30.555 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-09T17:53:30.568 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T17:53:30.568 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T17:53:30.583 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-09T17:53:30.583 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-09T17:53:30.625 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout:=========================================================================================== 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout:=========================================================================================== 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T17:53:30.631 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T17:53:30.632 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout:=========================================================================================== 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout:Remove 102 Packages 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 613 M 2026-03-09T17:53:30.633 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T17:53:30.659 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T17:53:30.659 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T17:53:30.675 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T17:53:30.675 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T17:53:30.690 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-09T17:53:30.690 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-09T17:53:30.773 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T17:53:30.773 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T17:53:30.820 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T17:53:30.820 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T17:53:30.828 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T17:53:30.838 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-09T17:53:30.838 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T17:53:30.846 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T17:53:30.848 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T17:53:30.848 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:30.848 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T17:53:30.848 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T17:53:30.848 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T17:53:30.848 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:30.848 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T17:53:30.862 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T17:53:30.866 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T17:53:30.866 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:30.866 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T17:53:30.866 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T17:53:30.866 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T17:53:30.866 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:30.867 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T17:53:30.881 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T17:53:30.885 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T17:53:30.885 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T17:53:30.904 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T17:53:30.905 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T17:53:30.935 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T17:53:30.935 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T17:53:30.939 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T17:53:30.943 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T17:53:30.948 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T17:53:30.952 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T17:53:30.952 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T17:53:30.961 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T17:53:30.964 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T17:53:30.966 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T17:53:30.966 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:30.966 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T17:53:30.966 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T17:53:30.966 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T17:53:30.966 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:30.966 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T17:53:30.972 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T17:53:30.972 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T17:53:30.976 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T17:53:30.977 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T17:53:30.977 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T17:53:30.981 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T17:53:30.985 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T17:53:30.989 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T17:53:30.990 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T17:53:30.996 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T17:53:31.000 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T17:53:31.006 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T17:53:31.006 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T17:53:31.008 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T17:53:31.008 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:31.008 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T17:53:31.008 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T17:53:31.008 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T17:53:31.008 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:31.009 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T17:53:31.013 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T17:53:31.013 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T17:53:31.022 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T17:53:31.035 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T17:53:31.035 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:31.035 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T17:53:31.035 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T17:53:31.035 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T17:53:31.035 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:31.037 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T17:53:31.037 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:31.037 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T17:53:31.037 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:31.041 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T17:53:31.046 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T17:53:31.051 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T17:53:31.056 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T17:53:31.058 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T17:53:31.063 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T17:53:31.067 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T17:53:31.068 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T17:53:31.068 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:31.068 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T17:53:31.068 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:31.069 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T17:53:31.076 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T17:53:31.077 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T17:53:31.079 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T17:53:31.084 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T17:53:31.084 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T17:53:31.087 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T17:53:31.089 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T17:53:31.089 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T17:53:31.094 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T17:53:31.096 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T17:53:31.098 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T17:53:31.099 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T17:53:31.106 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T17:53:31.108 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T17:53:31.114 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T17:53:31.127 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T17:53:31.129 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T17:53:31.134 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T17:53:31.136 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T17:53:31.140 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T17:53:31.140 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T17:53:31.150 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T17:53:31.156 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T17:53:31.159 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T17:53:31.165 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T17:53:31.165 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:31.165 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T17:53:31.165 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T17:53:31.165 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T17:53:31.165 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:31.171 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T17:53:31.188 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T17:53:31.231 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T17:53:31.245 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T17:53:31.248 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T17:53:31.250 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T17:53:31.251 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T17:53:31.257 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T17:53:31.271 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T17:53:31.271 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:31.271 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T17:53:31.271 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:31.281 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T17:53:31.288 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T17:53:31.343 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T17:53:31.343 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T17:53:31.348 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T17:53:31.352 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T17:53:31.352 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T17:53:31.354 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T17:53:31.355 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T17:53:31.359 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T17:53:31.365 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T17:53:31.365 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T17:53:31.375 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T17:53:31.389 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T17:53:31.395 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T17:53:31.405 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T17:53:31.412 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T17:53:31.444 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T17:53:31.454 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T17:53:31.463 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T17:53:31.606 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T17:53:31.619 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T17:53:31.619 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T17:53:31.619 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:31.620 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T17:53:31.631 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T17:53:31.634 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T17:53:31.636 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T17:53:31.646 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T17:53:31.649 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T17:53:31.649 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T17:53:31.649 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T17:53:31.649 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:31.650 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T17:53:31.657 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T17:53:31.657 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T17:53:31.670 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T17:53:31.688 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T17:53:31.689 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T17:53:31.692 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T17:53:31.695 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T17:53:31.702 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T17:53:31.716 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T17:53:31.717 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:31.717 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T17:53:31.717 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T17:53:31.717 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T17:53:31.717 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:31.718 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T17:53:31.721 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T17:53:31.727 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T17:53:31.729 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T17:53:31.731 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T17:53:31.732 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T17:53:31.736 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T17:53:31.738 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T17:53:31.741 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T17:53:31.744 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T17:53:31.747 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T17:53:31.752 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T17:53:31.752 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T17:53:31.752 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:31.752 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T17:53:31.752 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T17:53:31.752 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T17:53:31.752 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:31.753 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T17:53:31.757 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T17:53:31.764 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T17:53:31.767 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T17:53:31.769 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T17:53:31.772 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T17:53:31.774 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T17:53:31.778 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T17:53:31.782 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T17:53:31.787 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T17:53:31.798 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T17:53:31.806 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T17:53:31.813 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T17:53:31.817 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T17:53:31.820 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T17:53:31.825 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T17:53:31.827 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T17:53:31.830 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T17:53:31.830 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T17:53:31.830 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:31.830 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T17:53:31.831 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T17:53:31.833 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T17:53:31.836 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T17:53:31.848 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T17:53:31.851 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T17:53:31.853 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T17:53:31.853 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:31.853 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T17:53:31.853 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:31.854 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T17:53:31.856 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T17:53:31.859 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T17:53:31.859 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T17:53:31.862 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T17:53:31.863 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T17:53:31.864 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T17:53:31.865 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T17:53:31.866 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T17:53:31.868 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T17:53:31.871 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T17:53:31.873 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T17:53:31.875 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T17:53:31.877 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T17:53:31.878 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T17:53:31.881 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T17:53:31.884 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T17:53:31.887 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T17:53:31.887 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:31.887 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T17:53:31.887 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:31.887 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T17:53:31.887 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T17:53:31.888 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T17:53:31.889 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T17:53:31.892 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T17:53:31.893 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T17:53:31.895 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T17:53:31.895 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T17:53:31.897 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T17:53:31.897 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T17:53:31.900 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T17:53:31.901 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T17:53:31.903 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T17:53:31.905 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T17:53:31.905 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T17:53:31.908 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T17:53:31.910 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T17:53:31.911 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T17:53:31.914 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T17:53:31.915 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T17:53:31.918 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T17:53:31.918 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:31.918 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T17:53:31.918 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T17:53:31.918 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T17:53:31.918 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:31.918 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T17:53:31.919 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T17:53:31.922 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T17:53:31.924 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T17:53:31.927 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T17:53:31.928 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T17:53:31.930 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T17:53:31.931 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T17:53:31.932 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T17:53:31.932 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T17:53:31.935 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T17:53:31.935 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T17:53:31.935 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T17:53:31.937 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T17:53:31.940 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T17:53:31.940 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T17:53:31.940 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T17:53:31.942 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T17:53:31.944 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T17:53:31.945 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T17:53:31.946 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T17:53:31.947 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T17:53:31.950 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T17:53:31.951 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T17:53:31.954 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T17:53:31.955 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T17:53:31.956 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T17:53:31.959 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T17:53:31.961 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T17:53:31.963 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T17:53:31.964 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T17:53:31.966 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T17:53:31.967 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T17:53:31.968 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T17:53:31.969 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T17:53:31.974 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T17:53:31.974 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T17:53:31.978 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T17:53:31.978 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T17:53:31.981 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T17:53:31.989 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T17:53:31.994 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T17:53:31.998 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T17:53:32.000 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T17:53:32.002 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T17:53:32.008 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T17:53:32.009 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T17:53:32.011 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T17:53:32.022 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T17:53:32.022 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-09T17:53:32.022 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T17:53:32.022 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:32.025 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T17:53:32.028 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T17:53:32.033 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T17:53:32.034 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T17:53:32.035 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T17:53:32.039 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T17:53:32.043 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T17:53:32.062 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T17:53:32.062 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-09T17:53:32.062 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T17:53:32.062 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:32.069 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T17:53:32.070 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T17:53:32.070 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T17:53:32.070 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:32.070 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T17:53:32.071 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T17:53:32.071 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T17:53:32.071 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T17:53:32.080 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T17:53:32.085 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T17:53:32.108 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T17:53:32.108 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T17:53:32.111 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T17:53:32.113 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T17:53:32.114 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T17:53:32.118 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T17:53:32.119 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T17:53:32.121 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T17:53:32.121 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T17:53:32.121 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T17:53:32.124 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T17:53:32.127 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T17:53:32.128 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T17:53:32.128 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T17:53:32.135 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T17:53:32.140 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T17:53:32.142 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T17:53:32.144 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T17:53:32.145 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T17:53:32.147 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T17:53:32.149 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T17:53:32.152 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T17:53:32.153 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T17:53:32.155 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T17:53:32.155 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T17:53:32.157 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T17:53:32.163 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T17:53:32.168 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T17:53:32.175 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T17:53:32.178 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T17:53:32.183 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T17:53:32.186 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T17:53:32.192 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T17:53:32.196 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T17:53:32.199 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T17:53:32.209 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T17:53:32.215 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T17:53:32.220 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T17:53:32.223 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T17:53:32.225 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T17:53:32.232 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T17:53:32.237 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T17:53:32.242 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:32 vm05 systemd[1]: Stopping Ceph mon.c for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:53:32.242 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:53:32 vm05 systemd[1]: Stopping Ceph osd.6 for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:53:32.242 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:53:32 vm05 systemd[1]: Stopping Ceph osd.5 for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:53:32.243 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:53:32 vm05 systemd[1]: Stopping Ceph osd.7 for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:53:32.283 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:32 vm01 systemd[1]: Stopping Ceph mon.a for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:53:32.284 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:53:32 vm01 systemd[1]: Stopping Ceph mgr.x for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:53:32.284 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:53:32 vm01 systemd[1]: Stopping Ceph osd.0 for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:53:32.284 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:53:32 vm01 systemd[1]: Stopping Ceph osd.1 for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:53:32.284 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T17:53:32.284 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-09T17:53:32.284 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T17:53:32.284 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:32.295 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T17:53:32.336 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T17:53:32.336 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T17:53:32.359 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T17:53:32.367 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T17:53:32.371 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T17:53:32.373 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T17:53:32.373 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T17:53:32.501 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:32 vm05 systemd[1]: Stopping Ceph iscsi.iscsi.b for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:53:32.503 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:53:32 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-6[62565]: 2026-03-09T17:53:32.240+0000 7f6293d57640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.6 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T17:53:32.503 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:53:32 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-6[62565]: 2026-03-09T17:53:32.240+0000 7f6293d57640 -1 osd.6 60 *** Got signal Terminated *** 2026-03-09T17:53:32.504 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:53:32 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-6[62565]: 2026-03-09T17:53:32.240+0000 7f6293d57640 -1 osd.6 60 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T17:53:32.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:32 vm04 systemd[1]: Stopping Ceph mon.b for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:53:32.507 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:53:32 vm04 systemd[1]: Stopping Ceph osd.2 for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:53:32.507 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:53:32 vm04 systemd[1]: Stopping Ceph osd.3 for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:53:32.507 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:53:32 vm04 systemd[1]: Stopping Ceph osd.4 for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:53:32.549 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:32 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-a[52789]: 2026-03-09T17:53:32.479+0000 7f0af948a640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T17:53:32.549 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:32 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-a[52789]: 2026-03-09T17:53:32.479+0000 7f0af948a640 -1 mon.a@0(leader) e3 *** Got Signal Terminated *** 2026-03-09T17:53:32.549 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:53:32 vm01 podman[88304]: 2026-03-09 17:53:32.505519362 +0000 UTC m=+0.174592781 container died d63e78dabf111086fd87fb062b8a25fb399fa54b505f7721399461537f8fefdc (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-09T17:53:32.549 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:32 vm01 systemd[1]: Stopping Ceph iscsi.iscsi.a for 01455850-1bdf-11f1-910a-9936d43313cc... 2026-03-09T17:53:32.549 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:53:32 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-0[64363]: 2026-03-09T17:53:32.399+0000 7fe102431640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T17:53:32.549 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:53:32 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-0[64363]: 2026-03-09T17:53:32.399+0000 7fe102431640 -1 osd.0 60 *** Got signal Terminated *** 2026-03-09T17:53:32.549 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:53:32 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-0[64363]: 2026-03-09T17:53:32.399+0000 7fe102431640 -1 osd.0 60 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T17:53:32.550 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:53:32 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-1[69215]: 2026-03-09T17:53:32.440+0000 7fc08af69640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T17:53:32.550 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:53:32 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-1[69215]: 2026-03-09T17:53:32.440+0000 7fc08af69640 -1 osd.1 60 *** Got signal Terminated *** 2026-03-09T17:53:32.550 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:53:32 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-1[69215]: 2026-03-09T17:53:32.440+0000 7fc08af69640 -1 osd.1 60 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T17:53:32.757 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:32 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-b[53154]: 2026-03-09T17:53:32.615+0000 7f701b9c6640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T17:53:32.757 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:32 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-b[53154]: 2026-03-09T17:53:32.615+0000 7f701b9c6640 -1 mon.b@2(peon) e3 *** Got Signal Terminated *** 2026-03-09T17:53:32.757 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:32 vm04 podman[76621]: 2026-03-09 17:53:32.736952517 +0000 UTC m=+0.258587639 container died f68aca850be2509c7e97a7a668bcc501a0d4f4c9dd2de2bd7785fac32c40997d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-b, ceph=True, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, OSD_FLAVOR=default, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-09T17:53:32.758 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:53:32 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-2[56304]: 2026-03-09T17:53:32.558+0000 7f40338d5640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T17:53:32.758 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:53:32 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-2[56304]: 2026-03-09T17:53:32.558+0000 7f40338d5640 -1 osd.2 60 *** Got signal Terminated *** 2026-03-09T17:53:32.758 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:53:32 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-2[56304]: 2026-03-09T17:53:32.558+0000 7f40338d5640 -1 osd.2 60 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T17:53:32.758 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:53:32 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-4[66342]: 2026-03-09T17:53:32.595+0000 7f5bd66c2640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.4 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T17:53:32.758 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:53:32 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-4[66342]: 2026-03-09T17:53:32.595+0000 7f5bd66c2640 -1 osd.4 60 *** Got signal Terminated *** 2026-03-09T17:53:32.758 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:53:32 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-4[66342]: 2026-03-09T17:53:32.595+0000 7f5bd66c2640 -1 osd.4 60 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T17:53:32.758 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:53:32 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-3[61096]: 2026-03-09T17:53:32.555+0000 7f823cdc5640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.3 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T17:53:32.758 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:53:32 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-3[61096]: 2026-03-09T17:53:32.555+0000 7f823cdc5640 -1 osd.3 60 *** Got signal Terminated *** 2026-03-09T17:53:32.758 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:53:32 vm04 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-3[61096]: 2026-03-09T17:53:32.555+0000 7f823cdc5640 -1 osd.3 60 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T17:53:32.777 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:53:32 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-5[57405]: 2026-03-09T17:53:32.538+0000 7f706c173640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.5 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T17:53:32.777 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:53:32 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-5[57405]: 2026-03-09T17:53:32.538+0000 7f706c173640 -1 osd.5 60 *** Got signal Terminated *** 2026-03-09T17:53:32.777 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:53:32 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-5[57405]: 2026-03-09T17:53:32.538+0000 7f706c173640 -1 osd.5 60 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T17:53:32.779 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:32 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-c[53827]: 2026-03-09T17:53:32.528+0000 7f664c3fa640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.c -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T17:53:32.779 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:32 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-c[53827]: 2026-03-09T17:53:32.528+0000 7f664c3fa640 -1 mon.c@1(peon) e3 *** Got Signal Terminated *** 2026-03-09T17:53:32.779 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:32 vm05 podman[79849]: 2026-03-09 17:53:32.604106169 +0000 UTC m=+0.172969000 container died f158eba845cfcb82609246e5a51204cd11cd474b54d23b5a6d2adbc1fec52fa9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-c, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) 2026-03-09T17:53:32.779 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:32 vm05 podman[79849]: 2026-03-09 17:53:32.64577622 +0000 UTC m=+0.214639051 container remove f158eba845cfcb82609246e5a51204cd11cd474b54d23b5a6d2adbc1fec52fa9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-c, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, OSD_FLAVOR=default, ceph=True) 2026-03-09T17:53:32.779 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:32 vm05 bash[79849]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-c 2026-03-09T17:53:32.779 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:32 vm05 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.c.service: Deactivated successfully. 2026-03-09T17:53:32.779 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:32 vm05 systemd[1]: Stopped Ceph mon.c for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:53:32.779 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 09 17:53:32 vm05 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.c.service: Consumed 4.869s CPU time. 2026-03-09T17:53:32.779 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:53:32 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-7[67389]: 2026-03-09T17:53:32.565+0000 7f7527053640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.7 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T17:53:32.779 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:53:32 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-7[67389]: 2026-03-09T17:53:32.565+0000 7f7527053640 -1 osd.7 60 *** Got signal Terminated *** 2026-03-09T17:53:32.779 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:53:32 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-7[67389]: 2026-03-09T17:53:32.565+0000 7f7527053640 -1 osd.7 60 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T17:53:32.779 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:32 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug Shutdown received 2026-03-09T17:53:32.779 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:32 vm05 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b[70849]: debug No gateway configuration to remove on this host (vm05.local) 2026-03-09T17:53:32.779 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:32 vm05 podman[79863]: 2026-03-09 17:53:32.630129912 +0000 UTC m=+0.161185945 container died 49ceb8ca04352a3a76a1aef77660b8805d31d3594bacf7f93f961b0080d6b8cd (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3) 2026-03-09T17:53:32.779 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:32 vm05 podman[79863]: 2026-03-09 17:53:32.684443087 +0000 UTC m=+0.215499120 container remove 49ceb8ca04352a3a76a1aef77660b8805d31d3594bacf7f93f961b0080d6b8cd (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, org.label-schema.license=GPLv2) 2026-03-09T17:53:32.779 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:32 vm05 bash[79863]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-b 2026-03-09T17:53:32.811 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:32 vm01 podman[88319]: 2026-03-09 17:53:32.700411795 +0000 UTC m=+0.335382011 container died 98555f31d8dac30bb086772aab311a7e00bf9d59c18ac380dd9f5afca74e4afc (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-a, OSD_FLAVOR=default, io.buildah.version=1.41.3, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, CEPH_REF=squid, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T17:53:32.811 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:32 vm01 podman[88319]: 2026-03-09 17:53:32.752582854 +0000 UTC m=+0.387553080 container remove 98555f31d8dac30bb086772aab311a7e00bf9d59c18ac380dd9f5afca74e4afc (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-a, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.vendor=CentOS, CEPH_REF=squid, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, ceph=True) 2026-03-09T17:53:32.811 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:32 vm01 bash[88319]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-a 2026-03-09T17:53:32.811 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:32 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug Shutdown received 2026-03-09T17:53:32.811 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:32 vm01 ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a[73795]: debug No gateway configuration to remove on this host (vm01.local) 2026-03-09T17:53:32.811 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:32 vm01 podman[88382]: 2026-03-09 17:53:32.77976965 +0000 UTC m=+0.146926687 container died 966bf2fe12a9f8c64352f93ff81750ea31f506c92d9d0ba20c80a8bc938c27b9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, ceph=True, CEPH_REF=squid, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T17:53:32.811 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:53:32 vm01 podman[88304]: 2026-03-09 17:53:32.570796481 +0000 UTC m=+0.239869900 container remove d63e78dabf111086fd87fb062b8a25fb399fa54b505f7721399461537f8fefdc (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , ceph=True, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223) 2026-03-09T17:53:32.811 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:53:32 vm01 bash[88304]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-mgr-x 2026-03-09T17:53:32.811 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:53:32 vm01 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@mgr.x.service: Deactivated successfully. 2026-03-09T17:53:32.811 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:53:32 vm01 systemd[1]: Stopped Ceph mgr.x for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:53:32.811 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 09 17:53:32 vm01 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@mgr.x.service: Consumed 19.006s CPU time. 2026-03-09T17:53:33.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:32 vm04 podman[76621]: 2026-03-09 17:53:32.757436167 +0000 UTC m=+0.279071299 container remove f68aca850be2509c7e97a7a668bcc501a0d4f4c9dd2de2bd7785fac32c40997d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-b, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, org.label-schema.vendor=CentOS, ceph=True, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3) 2026-03-09T17:53:33.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:32 vm04 bash[76621]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-mon-b 2026-03-09T17:53:33.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:32 vm04 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.b.service: Deactivated successfully. 2026-03-09T17:53:33.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:32 vm04 systemd[1]: Stopped Ceph mon.b for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:53:33.041 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 09 17:53:32 vm04 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.b.service: Consumed 4.914s CPU time. 2026-03-09T17:53:33.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:32 vm05 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@iscsi.iscsi.b.service: Deactivated successfully. 2026-03-09T17:53:33.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:32 vm05 systemd[1]: Stopped Ceph iscsi.iscsi.b for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:53:33.070 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 09 17:53:32 vm05 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@iscsi.iscsi.b.service: Consumed 5.208s CPU time. 2026-03-09T17:53:33.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:32 vm01 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.a.service: Deactivated successfully. 2026-03-09T17:53:33.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:32 vm01 systemd[1]: Stopped Ceph mon.a for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:53:33.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 09 17:53:32 vm01 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.a.service: Consumed 6.620s CPU time. 2026-03-09T17:53:33.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:32 vm01 podman[88382]: 2026-03-09 17:53:32.811325063 +0000 UTC m=+0.178482109 container remove 966bf2fe12a9f8c64352f93ff81750ea31f506c92d9d0ba20c80a8bc938c27b9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, ceph=True, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-09T17:53:33.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:32 vm01 bash[88382]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-iscsi-iscsi-a 2026-03-09T17:53:33.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:32 vm01 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@iscsi.iscsi.a.service: Deactivated successfully. 2026-03-09T17:53:33.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:32 vm01 systemd[1]: Stopped Ceph iscsi.iscsi.a for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:53:33.227 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 09 17:53:32 vm01 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@iscsi.iscsi.a.service: Consumed 12.834s CPU time. 2026-03-09T17:53:37.559 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:53:37 vm05 podman[79850]: 2026-03-09 17:53:37.557685474 +0000 UTC m=+5.107025451 container died 01007efac186e6a3877bff7f046ca1d49a83f81cbfcee1b21971ec54920dfb36 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-5, ceph=True, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-09T17:53:37.559 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:53:37 vm05 podman[79775]: 2026-03-09 17:53:37.261041805 +0000 UTC m=+5.061246610 container died a7b407416475cac0a7c3cccf6396be0a15da0ca61253fc19405f140f2785b4e3 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-6, CEPH_REF=squid, org.label-schema.license=GPLv2, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223) 2026-03-09T17:53:37.559 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:53:37 vm05 podman[79775]: 2026-03-09 17:53:37.310538825 +0000 UTC m=+5.110743641 container remove a7b407416475cac0a7c3cccf6396be0a15da0ca61253fc19405f140f2785b4e3 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-6, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, OSD_FLAVOR=default) 2026-03-09T17:53:37.559 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:53:37 vm05 bash[79775]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-6 2026-03-09T17:53:37.559 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:53:37 vm05 bash[80007]: Error: statfs /var/log/ceph/01455850-1bdf-11f1-910a-9936d43313cc: no such file or directory 2026-03-09T17:53:37.560 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:53:37 vm05 podman[80007]: 2026-03-09 17:53:37.496333172 +0000 UTC m=+0.017221506 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T17:53:37.560 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:53:37 vm05 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.6.service: Deactivated successfully. 2026-03-09T17:53:37.560 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:53:37 vm05 systemd[1]: Stopped Ceph osd.6 for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:53:37.560 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 09 17:53:37 vm05 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.6.service: Consumed 2.467s CPU time. 2026-03-09T17:53:37.728 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:53:37 vm01 podman[88289]: 2026-03-09 17:53:37.445112272 +0000 UTC m=+5.135356245 container died f17d6a09c338b9b9e9e463ef1b85e5d90f3ae78db344cde40f95fc8dcc3fe50d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-0, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, ceph=True, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-09T17:53:37.728 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:53:37 vm01 podman[88289]: 2026-03-09 17:53:37.539364816 +0000 UTC m=+5.229608779 container remove f17d6a09c338b9b9e9e463ef1b85e5d90f3ae78db344cde40f95fc8dcc3fe50d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-0, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, ceph=True, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.build-date=20260223) 2026-03-09T17:53:37.728 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:53:37 vm01 bash[88289]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-0 2026-03-09T17:53:37.728 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:53:37 vm01 podman[88326]: 2026-03-09 17:53:37.492173408 +0000 UTC m=+5.091979978 container died 6b97853eb43b39fdf189a64bdf8f8a64f55dcf39ddd13eac9a1e9ec3e642e706 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-1, org.label-schema.license=GPLv2, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, org.label-schema.build-date=20260223) 2026-03-09T17:53:37.728 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:53:37 vm01 podman[88326]: 2026-03-09 17:53:37.527750207 +0000 UTC m=+5.127556777 container remove 6b97853eb43b39fdf189a64bdf8f8a64f55dcf39ddd13eac9a1e9ec3e642e706 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-1, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, OSD_FLAVOR=default, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T17:53:37.728 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:53:37 vm01 bash[88326]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-1 2026-03-09T17:53:37.814 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:53:37 vm05 podman[79850]: 2026-03-09 17:53:37.58508961 +0000 UTC m=+5.134429578 container remove 01007efac186e6a3877bff7f046ca1d49a83f81cbfcee1b21971ec54920dfb36 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-5, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, ceph=True, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223) 2026-03-09T17:53:37.814 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:53:37 vm05 bash[79850]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-5 2026-03-09T17:53:37.814 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:53:37 vm05 podman[79859]: 2026-03-09 17:53:37.596418556 +0000 UTC m=+5.142536372 container died 8a1ffa3d0d45c61f6a03fe5efc2ffb0e821952a85ff2b923357343c19649ec90 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-7, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, ceph=True, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T17:53:37.814 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:53:37 vm05 podman[79859]: 2026-03-09 17:53:37.630524406 +0000 UTC m=+5.176642212 container remove 8a1ffa3d0d45c61f6a03fe5efc2ffb0e821952a85ff2b923357343c19649ec90 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-7, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, CEPH_REF=squid, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, OSD_FLAVOR=default, org.label-schema.schema-version=1.0) 2026-03-09T17:53:37.814 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:53:37 vm05 bash[79859]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-7 2026-03-09T17:53:38.041 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:53:37 vm04 podman[76622]: 2026-03-09 17:53:37.634897157 +0000 UTC m=+5.161375748 container died 619ada5a151ab0b70242c58acb1b673969012e1fcbfeb5034cd41d7a84fb602c (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-4, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default) 2026-03-09T17:53:38.041 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:53:37 vm04 podman[76622]: 2026-03-09 17:53:37.713184215 +0000 UTC m=+5.239662806 container remove 619ada5a151ab0b70242c58acb1b673969012e1fcbfeb5034cd41d7a84fb602c (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-4, CEPH_REF=squid, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True) 2026-03-09T17:53:38.041 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:53:37 vm04 bash[76622]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-4 2026-03-09T17:53:38.041 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:53:37 vm04 podman[76607]: 2026-03-09 17:53:37.6163745 +0000 UTC m=+5.168777140 container died e6067fd11c048389bc87b1970f99f55e84c2d7d8d9343210ef9b70a8cfd90eb2 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-2, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , ceph=True, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-09T17:53:38.041 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:53:37 vm04 podman[76607]: 2026-03-09 17:53:37.745154266 +0000 UTC m=+5.297556906 container remove e6067fd11c048389bc87b1970f99f55e84c2d7d8d9343210ef9b70a8cfd90eb2 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-2, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, org.label-schema.schema-version=1.0) 2026-03-09T17:53:38.041 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:53:37 vm04 bash[76607]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-2 2026-03-09T17:53:38.041 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:53:37 vm04 podman[76609]: 2026-03-09 17:53:37.667755501 +0000 UTC m=+5.214558344 container died 114082d9a7884a0622778820ef4bb77d400a5b790b0693577c775ebc23817326 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-3, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T17:53:38.041 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:53:37 vm04 podman[76609]: 2026-03-09 17:53:37.701451203 +0000 UTC m=+5.248254046 container remove 114082d9a7884a0622778820ef4bb77d400a5b790b0693577c775ebc23817326 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-3, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T17:53:38.041 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:53:37 vm04 bash[76609]: ceph-01455850-1bdf-11f1-910a-9936d43313cc-osd-3 2026-03-09T17:53:38.071 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:53:37 vm05 bash[80106]: Error: statfs /var/log/ceph/01455850-1bdf-11f1-910a-9936d43313cc: no such file or directory 2026-03-09T17:53:38.071 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:53:37 vm05 podman[80106]: 2026-03-09 17:53:37.81006655 +0000 UTC m=+0.011666047 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T17:53:38.071 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:53:37 vm05 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.5.service: Deactivated successfully. 2026-03-09T17:53:38.071 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:53:37 vm05 systemd[1]: Stopped Ceph osd.5 for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:53:38.071 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 09 17:53:37 vm05 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.5.service: Consumed 2.332s CPU time. 2026-03-09T17:53:38.071 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:53:37 vm05 bash[80138]: Error: statfs /var/log/ceph/01455850-1bdf-11f1-910a-9936d43313cc: no such file or directory 2026-03-09T17:53:38.071 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:53:37 vm05 podman[80138]: 2026-03-09 17:53:37.890249058 +0000 UTC m=+0.009474605 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T17:53:38.071 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:53:37 vm05 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.7.service: Deactivated successfully. 2026-03-09T17:53:38.071 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:53:37 vm05 systemd[1]: Stopped Ceph osd.7 for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:53:38.071 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 09 17:53:37 vm05 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.7.service: Consumed 2.212s CPU time. 2026-03-09T17:53:38.228 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:53:37 vm01 bash[88586]: Error: statfs /var/log/ceph/01455850-1bdf-11f1-910a-9936d43313cc: no such file or directory 2026-03-09T17:53:38.228 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:53:37 vm01 podman[88586]: 2026-03-09 17:53:37.807973937 +0000 UTC m=+0.011152876 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T17:53:38.228 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:53:37 vm01 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.0.service: Deactivated successfully. 2026-03-09T17:53:38.228 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:53:37 vm01 systemd[1]: Stopped Ceph osd.0 for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:53:38.228 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 09 17:53:37 vm01 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.0.service: Consumed 2.623s CPU time. 2026-03-09T17:53:38.228 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:53:37 vm01 bash[88566]: Error: statfs /var/log/ceph/01455850-1bdf-11f1-910a-9936d43313cc: no such file or directory 2026-03-09T17:53:38.228 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:53:37 vm01 podman[88566]: 2026-03-09 17:53:37.784283094 +0000 UTC m=+0.022611233 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T17:53:38.228 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:53:37 vm01 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.1.service: Deactivated successfully. 2026-03-09T17:53:38.228 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:53:37 vm01 systemd[1]: Stopped Ceph osd.1 for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:53:38.228 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 09 17:53:37 vm01 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.1.service: Consumed 2.651s CPU time. 2026-03-09T17:53:38.371 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T17:53:38.372 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /sys 2026-03-09T17:53:38.372 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /proc 2026-03-09T17:53:38.372 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /mnt 2026-03-09T17:53:38.372 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /var/tmp 2026-03-09T17:53:38.372 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /home 2026-03-09T17:53:38.372 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /root 2026-03-09T17:53:38.372 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /tmp 2026-03-09T17:53:38.372 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:38.382 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T17:53:38.383 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T17:53:38.383 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /sys 2026-03-09T17:53:38.383 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /proc 2026-03-09T17:53:38.383 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /mnt 2026-03-09T17:53:38.383 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /var/tmp 2026-03-09T17:53:38.383 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /home 2026-03-09T17:53:38.383 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /root 2026-03-09T17:53:38.383 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /tmp 2026-03-09T17:53:38.383 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:38.391 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T17:53:38.399 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T17:53:38.399 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T17:53:38.407 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T17:53:38.409 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T17:53:38.410 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T17:53:38.410 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T17:53:38.412 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T17:53:38.414 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T17:53:38.416 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T17:53:38.416 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T17:53:38.417 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T17:53:38.420 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T17:53:38.422 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T17:53:38.425 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T17:53:38.427 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T17:53:38.427 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T17:53:38.431 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T17:53:38.433 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T17:53:38.435 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T17:53:38.438 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T17:53:38.440 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T17:53:38.441 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T17:53:38.443 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T17:53:38.445 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T17:53:38.446 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T17:53:38.448 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T17:53:38.450 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T17:53:38.453 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T17:53:38.455 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T17:53:38.457 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T17:53:38.457 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T17:53:38.464 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T17:53:38.469 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T17:53:38.469 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T17:53:38.541 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T17:53:38.541 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /sys 2026-03-09T17:53:38.541 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /proc 2026-03-09T17:53:38.541 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /mnt 2026-03-09T17:53:38.541 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /var/tmp 2026-03-09T17:53:38.541 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /home 2026-03-09T17:53:38.541 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /root 2026-03-09T17:53:38.541 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /tmp 2026-03-09T17:53:38.541 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:38.541 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:53:38 vm04 bash[76839]: Error: statfs /var/log/ceph/01455850-1bdf-11f1-910a-9936d43313cc: no such file or directory 2026-03-09T17:53:38.541 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:53:38 vm04 podman[76839]: 2026-03-09 17:53:38.102128641 +0000 UTC m=+0.037641713 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T17:53:38.541 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:53:38 vm04 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.2.service: Deactivated successfully. 2026-03-09T17:53:38.541 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:53:38 vm04 systemd[1]: Stopped Ceph osd.2 for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:53:38.541 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 09 17:53:38 vm04 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.2.service: Consumed 3.200s CPU time. 2026-03-09T17:53:38.541 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:53:38 vm04 bash[76843]: Error: statfs /var/log/ceph/01455850-1bdf-11f1-910a-9936d43313cc: no such file or directory 2026-03-09T17:53:38.541 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:53:38 vm04 podman[76843]: 2026-03-09 17:53:38.09664927 +0000 UTC m=+0.027405032 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T17:53:38.541 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:53:38 vm04 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.3.service: Deactivated successfully. 2026-03-09T17:53:38.541 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:53:38 vm04 systemd[1]: Stopped Ceph osd.3 for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:53:38.541 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 09 17:53:38 vm04 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.3.service: Consumed 3.737s CPU time. 2026-03-09T17:53:38.541 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:53:38 vm04 bash[76842]: Error: statfs /var/log/ceph/01455850-1bdf-11f1-910a-9936d43313cc: no such file or directory 2026-03-09T17:53:38.541 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:53:38 vm04 podman[76842]: 2026-03-09 17:53:38.094152394 +0000 UTC m=+0.022262602 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T17:53:38.541 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:53:38 vm04 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.4.service: Deactivated successfully. 2026-03-09T17:53:38.541 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:53:38 vm04 systemd[1]: Stopped Ceph osd.4 for 01455850-1bdf-11f1-910a-9936d43313cc. 2026-03-09T17:53:38.542 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 09 17:53:38 vm04 systemd[1]: ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.4.service: Consumed 2.255s CPU time. 2026-03-09T17:53:38.550 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T17:53:38.564 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T17:53:38.565 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T17:53:38.566 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T17:53:38.570 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T17:53:38.570 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T17:53:38.570 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T17:53:38.570 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T17:53:38.570 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T17:53:38.570 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T17:53:38.570 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T17:53:38.570 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T17:53:38.571 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T17:53:38.571 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T17:53:38.571 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T17:53:38.571 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T17:53:38.571 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T17:53:38.573 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T17:53:38.574 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T17:53:38.579 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T17:53:38.582 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T17:53:38.585 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T17:53:38.587 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T17:53:38.589 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T17:53:38.589 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T17:53:38.604 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T17:53:38.607 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T17:53:38.609 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T17:53:38.612 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T17:53:38.615 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T17:53:38.620 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T17:53:38.629 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T17:53:38.633 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T17:53:38.633 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T17:53:38.644 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T17:53:38.644 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:38.644 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-09T17:53:38.644 INFO:teuthology.orchestra.run.vm05.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T17:53:38.644 INFO:teuthology.orchestra.run.vm05.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T17:53:38.645 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:38.646 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T17:53:38.649 INFO:teuthology.orchestra.run.vm01.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T17:53:38.650 INFO:teuthology.orchestra.run.vm01.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:38.651 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:38.743 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T17:53:38.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T17:53:38.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T17:53:38.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T17:53:38.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T17:53:38.743 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T17:53:38.744 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T17:53:38.744 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T17:53:38.744 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T17:53:38.744 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T17:53:38.744 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T17:53:38.744 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T17:53:38.744 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T17:53:38.744 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T17:53:38.744 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T17:53:38.744 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T17:53:38.745 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T17:53:38.745 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T17:53:38.745 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T17:53:38.745 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T17:53:38.745 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T17:53:38.745 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T17:53:38.745 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T17:53:38.745 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T17:53:38.745 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T17:53:38.745 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T17:53:38.745 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T17:53:38.745 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T17:53:38.745 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T17:53:38.746 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T17:53:38.746 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T17:53:38.746 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T17:53:38.746 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T17:53:38.746 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T17:53:38.746 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T17:53:38.746 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T17:53:38.746 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T17:53:38.746 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T17:53:38.746 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T17:53:38.746 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T17:53:38.746 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T17:53:38.746 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T17:53:38.747 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T17:53:38.748 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T17:53:38.749 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T17:53:38.835 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T17:53:38.835 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T17:53:38.836 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:38.837 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:38.864 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:38.864 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:38.865 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-09T17:53:38.865 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:38.865 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-09T17:53:38.865 INFO:teuthology.orchestra.run.vm05.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T17:53:38.865 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:38.865 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-09T17:53:38.865 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:38.865 INFO:teuthology.orchestra.run.vm05.stdout:Remove 1 Package 2026-03-09T17:53:38.865 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:38.865 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 775 k 2026-03-09T17:53:38.865 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-09T17:53:38.867 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-09T17:53:38.867 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm01.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm01.stdout:Remove 1 Package 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 775 k 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-09T17:53:38.868 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-09T17:53:38.870 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T17:53:38.870 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T17:53:38.871 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T17:53:38.872 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T17:53:38.885 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-09T17:53:38.885 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T17:53:38.889 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T17:53:38.889 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T17:53:38.996 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T17:53:38.998 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T17:53:39.038 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T17:53:39.038 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:39.038 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T17:53:39.038 INFO:teuthology.orchestra.run.vm01.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:39.038 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:39.038 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:39.040 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T17:53:39.040 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:39.040 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-09T17:53:39.040 INFO:teuthology.orchestra.run.vm05.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:39.040 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:39.040 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:39.061 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:39.061 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:39.061 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T17:53:39.061 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:39.061 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T17:53:39.061 INFO:teuthology.orchestra.run.vm04.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T17:53:39.061 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:39.061 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T17:53:39.061 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:39.061 INFO:teuthology.orchestra.run.vm04.stdout:Remove 1 Package 2026-03-09T17:53:39.061 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:39.061 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 775 k 2026-03-09T17:53:39.061 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T17:53:39.064 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T17:53:39.064 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T17:53:39.065 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T17:53:39.066 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T17:53:39.083 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T17:53:39.084 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T17:53:39.199 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T17:53:39.229 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T17:53:39.230 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:39.233 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:39.234 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:39.234 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:39.237 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T17:53:39.237 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:39.241 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:39.241 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:39.241 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:39.250 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T17:53:39.251 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:39.251 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T17:53:39.251 INFO:teuthology.orchestra.run.vm04.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T17:53:39.251 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:39.251 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:39.405 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr 2026-03-09T17:53:39.405 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:39.408 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:39.409 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:39.409 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:39.412 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: ceph-mgr 2026-03-09T17:53:39.413 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:39.416 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:39.417 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:39.417 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:39.444 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T17:53:39.444 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:39.447 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:39.448 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:39.448 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:39.578 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T17:53:39.578 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:39.581 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:39.582 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:39.582 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:39.589 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T17:53:39.589 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:39.593 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:39.594 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:39.594 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:39.631 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr 2026-03-09T17:53:39.631 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:39.635 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:39.635 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:39.635 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:39.748 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T17:53:39.748 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:39.751 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:39.752 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:39.752 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:39.760 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T17:53:39.761 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:39.764 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:39.764 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:39.765 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:39.807 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T17:53:39.807 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:39.810 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:39.811 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:39.811 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:39.917 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-rook 2026-03-09T17:53:39.917 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:39.920 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:39.921 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:39.921 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:39.929 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: ceph-mgr-rook 2026-03-09T17:53:39.929 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:39.933 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:39.933 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:39.933 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:39.984 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T17:53:39.984 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:39.987 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:39.988 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:39.988 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:40.089 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T17:53:40.089 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:40.092 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:40.093 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:40.093 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:40.105 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T17:53:40.105 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:40.109 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:40.109 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:40.109 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:40.161 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-rook 2026-03-09T17:53:40.161 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:40.165 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:40.165 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:40.166 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:40.272 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:40.273 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:40.273 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T17:53:40.273 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:40.273 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T17:53:40.273 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T17:53:40.273 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:40.273 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T17:53:40.273 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:40.273 INFO:teuthology.orchestra.run.vm01.stdout:Remove 1 Package 2026-03-09T17:53:40.273 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:40.273 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 3.6 M 2026-03-09T17:53:40.273 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T17:53:40.275 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T17:53:40.275 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T17:53:40.284 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T17:53:40.284 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T17:53:40.288 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:40.288 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:40.289 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-09T17:53:40.289 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:40.289 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-09T17:53:40.289 INFO:teuthology.orchestra.run.vm05.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T17:53:40.289 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:40.289 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-09T17:53:40.289 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:40.289 INFO:teuthology.orchestra.run.vm05.stdout:Remove 1 Package 2026-03-09T17:53:40.289 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:40.289 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 3.6 M 2026-03-09T17:53:40.289 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-09T17:53:40.290 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-09T17:53:40.291 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-09T17:53:40.300 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-09T17:53:40.300 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-09T17:53:40.310 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T17:53:40.324 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T17:53:40.326 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-09T17:53:40.339 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T17:53:40.339 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:40.340 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T17:53:40.343 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:40.343 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:40.343 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:40.408 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T17:53:40.411 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T17:53:40.548 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:40.548 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:40.548 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T17:53:40.549 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:40.549 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T17:53:40.549 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T17:53:40.549 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:40.549 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T17:53:40.549 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:40.549 INFO:teuthology.orchestra.run.vm04.stdout:Remove 1 Package 2026-03-09T17:53:40.549 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:40.549 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 3.6 M 2026-03-09T17:53:40.549 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T17:53:40.550 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T17:53:40.550 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T17:53:40.560 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T17:53:40.560 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T17:53:40.568 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T17:53:40.568 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:40.568 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T17:53:40.568 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:40.568 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:40.568 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:40.569 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T17:53:40.569 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:40.569 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-09T17:53:40.569 INFO:teuthology.orchestra.run.vm05.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:40.569 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:40.569 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:40.591 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T17:53:40.605 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T17:53:40.685 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T17:53:40.739 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T17:53:40.740 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:40.740 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T17:53:40.740 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:40.740 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:40.740 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:40.753 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-volume 2026-03-09T17:53:40.753 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:40.755 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:40.756 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:40.756 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:40.763 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: ceph-volume 2026-03-09T17:53:40.763 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:40.765 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:40.766 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:40.766 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:40.951 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-volume 2026-03-09T17:53:40.951 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:40.954 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:40.954 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:40.954 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repo Size 2026-03-09T17:53:40.954 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:40.954 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T17:53:40.954 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T17:53:40.954 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-09T17:53:40.954 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T17:53:40.955 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:40.955 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T17:53:40.955 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:40.955 INFO:teuthology.orchestra.run.vm01.stdout:Remove 2 Packages 2026-03-09T17:53:40.955 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:40.955 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:40.955 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 610 k 2026-03-09T17:53:40.955 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T17:53:40.955 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:40.955 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:40.956 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T17:53:40.957 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repo Size 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout:Removing dependent packages: 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout:Remove 2 Packages 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 610 k 2026-03-09T17:53:40.960 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-09T17:53:40.962 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-09T17:53:40.962 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-09T17:53:40.967 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T17:53:40.967 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T17:53:40.973 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-09T17:53:40.973 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-09T17:53:40.993 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T17:53:40.995 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:40.998 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-09T17:53:41.000 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:41.008 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T17:53:41.012 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T17:53:41.071 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T17:53:41.072 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:41.086 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T17:53:41.086 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:41.125 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T17:53:41.125 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:41.125 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T17:53:41.125 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.125 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.125 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:41.125 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:41.140 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T17:53:41.140 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:41.140 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-09T17:53:41.140 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.140 INFO:teuthology.orchestra.run.vm05.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.140 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:41.140 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:41.158 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:41.159 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:41.159 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repo Size 2026-03-09T17:53:41.159 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:41.159 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T17:53:41.159 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T17:53:41.159 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-09T17:53:41.159 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T17:53:41.159 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:41.159 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T17:53:41.159 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:41.159 INFO:teuthology.orchestra.run.vm04.stdout:Remove 2 Packages 2026-03-09T17:53:41.159 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:41.159 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 610 k 2026-03-09T17:53:41.159 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T17:53:41.161 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T17:53:41.161 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T17:53:41.174 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T17:53:41.174 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T17:53:41.201 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T17:53:41.204 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:41.217 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T17:53:41.283 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T17:53:41.283 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T17:53:41.322 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repo Size 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout:Remove 3 Packages 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 3.7 M 2026-03-09T17:53:41.323 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T17:53:41.325 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T17:53:41.325 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T17:53:41.329 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T17:53:41.329 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:41.329 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T17:53:41.329 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.329 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.329 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:41.329 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:41.338 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repo Size 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout:Removing dependent packages: 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout:Removing unused dependencies: 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout:Remove 3 Packages 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 3.7 M 2026-03-09T17:53:41.339 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-09T17:53:41.341 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-09T17:53:41.341 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-09T17:53:41.341 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T17:53:41.342 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T17:53:41.358 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-09T17:53:41.358 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-09T17:53:41.376 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T17:53:41.378 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T17:53:41.382 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T17:53:41.382 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T17:53:41.389 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-09T17:53:41.392 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T17:53:41.393 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T17:53:41.393 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T17:53:41.444 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T17:53:41.444 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T17:53:41.444 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T17:53:41.456 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T17:53:41.456 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T17:53:41.456 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T17:53:41.485 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T17:53:41.485 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:41.485 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T17:53:41.485 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.485 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.485 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.485 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:41.485 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:41.495 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T17:53:41.495 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:41.495 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-09T17:53:41.495 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.495 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.495 INFO:teuthology.orchestra.run.vm05.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.495 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:41.495 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:41.539 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repo Size 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout:Remove 3 Packages 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:41.540 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 3.7 M 2026-03-09T17:53:41.541 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T17:53:41.542 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T17:53:41.543 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T17:53:41.561 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T17:53:41.562 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T17:53:41.594 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T17:53:41.596 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T17:53:41.598 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T17:53:41.598 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T17:53:41.659 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: libcephfs-devel 2026-03-09T17:53:41.659 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:41.662 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:41.663 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:41.663 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:41.664 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T17:53:41.664 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T17:53:41.665 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T17:53:41.669 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: libcephfs-devel 2026-03-09T17:53:41.669 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:41.671 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:41.672 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:41.672 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:41.703 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T17:53:41.703 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:41.703 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T17:53:41.703 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.703 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.703 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:41.703 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:41.703 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:41.832 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout:Remove 20 Packages 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 79 M 2026-03-09T17:53:41.834 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-09T17:53:41.838 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-09T17:53:41.838 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-09T17:53:41.844 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:41.845 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:41.845 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-09T17:53:41.845 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:41.845 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-09T17:53:41.845 INFO:teuthology.orchestra.run.vm05.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T17:53:41.845 INFO:teuthology.orchestra.run.vm05.stdout:Removing dependent packages: 2026-03-09T17:53:41.845 INFO:teuthology.orchestra.run.vm05.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T17:53:41.845 INFO:teuthology.orchestra.run.vm05.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T17:53:41.845 INFO:teuthology.orchestra.run.vm05.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T17:53:41.845 INFO:teuthology.orchestra.run.vm05.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T17:53:41.845 INFO:teuthology.orchestra.run.vm05.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T17:53:41.845 INFO:teuthology.orchestra.run.vm05.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T17:53:41.845 INFO:teuthology.orchestra.run.vm05.stdout:Removing unused dependencies: 2026-03-09T17:53:41.845 INFO:teuthology.orchestra.run.vm05.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout:Remove 20 Packages 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 79 M 2026-03-09T17:53:41.846 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-09T17:53:41.849 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-09T17:53:41.849 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-09T17:53:41.858 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-09T17:53:41.858 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-09T17:53:41.871 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: libcephfs-devel 2026-03-09T17:53:41.871 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:41.871 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-09T17:53:41.871 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-09T17:53:41.874 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:41.874 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:41.874 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:41.899 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-09T17:53:41.902 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T17:53:41.904 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T17:53:41.907 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T17:53:41.907 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T17:53:41.911 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-09T17:53:41.913 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T17:53:41.915 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T17:53:41.918 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T17:53:41.918 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T17:53:41.919 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T17:53:41.921 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T17:53:41.923 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T17:53:41.924 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T17:53:41.926 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T17:53:41.928 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T17:53:41.928 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T17:53:41.932 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T17:53:41.934 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T17:53:41.935 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T17:53:41.937 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T17:53:41.938 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T17:53:41.941 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T17:53:41.941 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T17:53:41.941 INFO:teuthology.orchestra.run.vm01.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T17:53:41.941 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:41.941 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T17:53:41.941 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T17:53:41.954 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T17:53:41.955 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T17:53:41.955 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T17:53:41.955 INFO:teuthology.orchestra.run.vm05.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T17:53:41.955 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:41.956 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T17:53:41.959 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T17:53:41.962 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T17:53:41.965 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T17:53:41.968 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T17:53:41.968 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T17:53:41.969 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T17:53:41.970 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T17:53:41.971 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T17:53:41.973 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T17:53:41.973 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T17:53:41.976 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T17:53:41.979 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T17:53:41.982 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T17:53:41.984 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T17:53:41.985 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T17:53:41.986 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T17:53:41.987 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T17:53:42.001 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T17:53:42.051 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T17:53:42.051 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T17:53:42.051 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T17:53:42.051 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T17:53:42.051 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T17:53:42.051 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T17:53:42.051 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T17:53:42.051 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T17:53:42.051 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T17:53:42.051 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T17:53:42.051 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T17:53:42.051 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T17:53:42.051 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T17:53:42.051 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T17:53:42.052 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T17:53:42.052 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T17:53:42.052 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T17:53:42.052 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T17:53:42.052 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T17:53:42.052 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T17:53:42.061 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout:Remove 20 Packages 2026-03-09T17:53:42.063 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 79 M 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T17:53:42.064 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T17:53:42.067 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-09T17:53:42.067 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-09T17:53:42.090 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-09T17:53:42.091 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-09T17:53:42.096 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-09T17:53:42.107 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:42.133 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-09T17:53:42.136 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T17:53:42.139 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T17:53:42.143 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T17:53:42.143 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T17:53:42.158 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T17:53:42.160 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T17:53:42.162 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T17:53:42.164 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T17:53:42.166 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T17:53:42.168 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T17:53:42.169 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T17:53:42.184 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T17:53:42.185 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T17:53:42.185 INFO:teuthology.orchestra.run.vm04.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T17:53:42.185 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:42.200 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T17:53:42.202 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T17:53:42.206 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T17:53:42.210 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T17:53:42.212 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T17:53:42.215 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T17:53:42.217 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T17:53:42.219 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T17:53:42.221 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T17:53:42.235 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T17:53:42.287 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: librbd1 2026-03-09T17:53:42.287 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:42.290 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:42.290 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:42.290 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T17:53:42.304 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T17:53:42.308 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: librbd1 2026-03-09T17:53:42.308 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:42.310 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:42.311 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:42.311 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:42.353 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T17:53:42.353 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:42.353 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-09T17:53:42.353 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-09T17:53:42.354 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:42.468 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rados 2026-03-09T17:53:42.468 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:42.470 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:42.470 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:42.471 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:42.505 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: python3-rados 2026-03-09T17:53:42.505 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:42.508 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:42.509 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:42.509 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:42.570 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: librbd1 2026-03-09T17:53:42.570 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:42.574 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:42.575 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:42.575 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:42.643 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rgw 2026-03-09T17:53:42.643 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:42.645 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:42.646 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:42.646 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:42.694 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: python3-rgw 2026-03-09T17:53:42.695 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:42.697 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:42.698 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:42.698 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:42.771 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rados 2026-03-09T17:53:42.771 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:42.773 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:42.774 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:42.774 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:42.832 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-cephfs 2026-03-09T17:53:42.832 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:42.834 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:42.835 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:42.835 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:42.868 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: python3-cephfs 2026-03-09T17:53:42.868 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:42.871 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:42.871 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:42.871 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:42.956 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rgw 2026-03-09T17:53:42.956 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:42.958 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:42.959 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:42.959 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:43.007 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rbd 2026-03-09T17:53:43.007 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:43.009 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:43.010 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:43.010 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:43.036 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: python3-rbd 2026-03-09T17:53:43.036 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:43.038 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:43.039 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:43.039 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:43.137 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-cephfs 2026-03-09T17:53:43.137 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:43.140 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:43.140 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:43.140 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:43.179 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-fuse 2026-03-09T17:53:43.179 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:43.181 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:43.182 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:43.182 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:43.209 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: rbd-fuse 2026-03-09T17:53:43.209 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:43.211 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:43.212 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:43.212 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:43.326 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rbd 2026-03-09T17:53:43.326 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:43.329 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:43.329 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:43.329 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:43.360 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-mirror 2026-03-09T17:53:43.360 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:43.362 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:43.363 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:43.363 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:43.381 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: rbd-mirror 2026-03-09T17:53:43.381 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:43.384 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:43.385 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:43.385 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:43.519 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-fuse 2026-03-09T17:53:43.519 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:43.522 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:43.523 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:43.523 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:43.548 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-nbd 2026-03-09T17:53:43.548 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-09T17:53:43.551 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-09T17:53:43.552 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-09T17:53:43.552 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-09T17:53:43.563 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: rbd-nbd 2026-03-09T17:53:43.563 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-09T17:53:43.565 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-09T17:53:43.566 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-09T17:53:43.566 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-09T17:53:43.578 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean all 2026-03-09T17:53:43.590 DEBUG:teuthology.orchestra.run.vm05:> sudo yum clean all 2026-03-09T17:53:43.715 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-mirror 2026-03-09T17:53:43.715 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:43.716 INFO:teuthology.orchestra.run.vm01.stdout:56 files removed 2026-03-09T17:53:43.718 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:43.718 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:43.718 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:43.728 INFO:teuthology.orchestra.run.vm05.stdout:56 files removed 2026-03-09T17:53:43.740 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T17:53:43.757 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T17:53:43.765 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean expire-cache 2026-03-09T17:53:43.785 DEBUG:teuthology.orchestra.run.vm05:> sudo yum clean expire-cache 2026-03-09T17:53:43.897 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-nbd 2026-03-09T17:53:43.897 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-09T17:53:43.900 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-09T17:53:43.900 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-09T17:53:43.900 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-09T17:53:43.922 INFO:teuthology.orchestra.run.vm01.stdout:Cache was expired 2026-03-09T17:53:43.923 INFO:teuthology.orchestra.run.vm01.stdout:0 files removed 2026-03-09T17:53:43.925 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean all 2026-03-09T17:53:43.942 INFO:teuthology.orchestra.run.vm05.stdout:Cache was expired 2026-03-09T17:53:43.943 INFO:teuthology.orchestra.run.vm05.stdout:0 files removed 2026-03-09T17:53:43.944 DEBUG:teuthology.parallel:result is None 2026-03-09T17:53:43.963 DEBUG:teuthology.parallel:result is None 2026-03-09T17:53:44.055 INFO:teuthology.orchestra.run.vm04.stdout:56 files removed 2026-03-09T17:53:44.077 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T17:53:44.102 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean expire-cache 2026-03-09T17:53:44.256 INFO:teuthology.orchestra.run.vm04.stdout:Cache was expired 2026-03-09T17:53:44.256 INFO:teuthology.orchestra.run.vm04.stdout:0 files removed 2026-03-09T17:53:44.274 DEBUG:teuthology.parallel:result is None 2026-03-09T17:53:44.274 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm01.local 2026-03-09T17:53:44.274 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm04.local 2026-03-09T17:53:44.274 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm05.local 2026-03-09T17:53:44.274 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T17:53:44.275 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T17:53:44.275 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T17:53:44.299 DEBUG:teuthology.orchestra.run.vm01:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T17:53:44.299 DEBUG:teuthology.orchestra.run.vm05:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T17:53:44.301 DEBUG:teuthology.orchestra.run.vm04:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T17:53:44.364 DEBUG:teuthology.parallel:result is None 2026-03-09T17:53:44.369 DEBUG:teuthology.parallel:result is None 2026-03-09T17:53:44.370 DEBUG:teuthology.parallel:result is None 2026-03-09T17:53:44.370 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-09T17:53:44.374 INFO:tasks.cephadm:Teardown begin 2026-03-09T17:53:44.374 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T17:53:44.427 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T17:53:44.455 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T17:53:44.480 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-09T17:53:44.480 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01455850-1bdf-11f1-910a-9936d43313cc -- ceph mgr module disable cephadm 2026-03-09T17:53:44.644 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/mon.a/config 2026-03-09T17:53:44.660 INFO:teuthology.orchestra.run.vm01.stderr:Error: statfs /etc/ceph/ceph.client.admin.keyring: no such file or directory 2026-03-09T17:53:44.683 DEBUG:teuthology.orchestra.run:got remote process result: 125 2026-03-09T17:53:44.683 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-09T17:53:44.683 DEBUG:teuthology.orchestra.run.vm01:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T17:53:44.700 DEBUG:teuthology.orchestra.run.vm04:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T17:53:44.719 DEBUG:teuthology.orchestra.run.vm05:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T17:53:44.736 INFO:tasks.cephadm:Stopping all daemons... 2026-03-09T17:53:44.736 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-09T17:53:44.736 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.a 2026-03-09T17:53:44.773 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.a.service' 2026-03-09T17:53:44.847 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T17:53:44.847 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-09T17:53:44.847 INFO:tasks.cephadm.mon.c:Stopping mon.b... 2026-03-09T17:53:44.847 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.b 2026-03-09T17:53:44.884 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.b.service' 2026-03-09T17:53:44.966 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T17:53:44.966 INFO:tasks.cephadm.mon.c:Stopped mon.b 2026-03-09T17:53:44.966 INFO:tasks.cephadm.mon.c:Stopping mon.c... 2026-03-09T17:53:44.966 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.c 2026-03-09T17:53:44.998 DEBUG:teuthology.orchestra.run.vm05:> sudo pkill -f 'journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@mon.c.service' 2026-03-09T17:53:45.069 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T17:53:45.069 INFO:tasks.cephadm.mon.c:Stopped mon.c 2026-03-09T17:53:45.069 INFO:tasks.cephadm.mgr.x:Stopping mgr.x... 2026-03-09T17:53:45.069 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-01455850-1bdf-11f1-910a-9936d43313cc@mgr.x 2026-03-09T17:53:45.101 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@mgr.x.service' 2026-03-09T17:53:45.173 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T17:53:45.173 INFO:tasks.cephadm.mgr.x:Stopped mgr.x 2026-03-09T17:53:45.173 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-09T17:53:45.173 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.0 2026-03-09T17:53:45.243 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.0.service' 2026-03-09T17:53:45.312 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T17:53:45.312 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-09T17:53:45.312 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-09T17:53:45.312 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.1 2026-03-09T17:53:45.380 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.1.service' 2026-03-09T17:53:45.449 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T17:53:45.449 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-09T17:53:45.450 INFO:tasks.cephadm.osd.2:Stopping osd.2... 2026-03-09T17:53:45.450 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.2 2026-03-09T17:53:45.482 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.2.service' 2026-03-09T17:53:45.551 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T17:53:45.551 INFO:tasks.cephadm.osd.2:Stopped osd.2 2026-03-09T17:53:45.551 INFO:tasks.cephadm.osd.3:Stopping osd.3... 2026-03-09T17:53:45.552 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.3 2026-03-09T17:53:45.621 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.3.service' 2026-03-09T17:53:45.691 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T17:53:45.691 INFO:tasks.cephadm.osd.3:Stopped osd.3 2026-03-09T17:53:45.691 INFO:tasks.cephadm.osd.4:Stopping osd.4... 2026-03-09T17:53:45.691 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.4 2026-03-09T17:53:45.761 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.4.service' 2026-03-09T17:53:45.830 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T17:53:45.830 INFO:tasks.cephadm.osd.4:Stopped osd.4 2026-03-09T17:53:45.830 INFO:tasks.cephadm.osd.5:Stopping osd.5... 2026-03-09T17:53:45.830 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.5 2026-03-09T17:53:45.861 DEBUG:teuthology.orchestra.run.vm05:> sudo pkill -f 'journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.5.service' 2026-03-09T17:53:45.932 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T17:53:45.932 INFO:tasks.cephadm.osd.5:Stopped osd.5 2026-03-09T17:53:45.932 INFO:tasks.cephadm.osd.6:Stopping osd.6... 2026-03-09T17:53:45.932 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.6 2026-03-09T17:53:46.001 DEBUG:teuthology.orchestra.run.vm05:> sudo pkill -f 'journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.6.service' 2026-03-09T17:53:46.071 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T17:53:46.071 INFO:tasks.cephadm.osd.6:Stopped osd.6 2026-03-09T17:53:46.071 INFO:tasks.cephadm.osd.7:Stopping osd.7... 2026-03-09T17:53:46.071 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.7 2026-03-09T17:53:46.138 DEBUG:teuthology.orchestra.run.vm05:> sudo pkill -f 'journalctl -f -n 0 -u ceph-01455850-1bdf-11f1-910a-9936d43313cc@osd.7.service' 2026-03-09T17:53:46.205 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T17:53:46.205 INFO:tasks.cephadm.osd.7:Stopped osd.7 2026-03-09T17:53:46.205 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 01455850-1bdf-11f1-910a-9936d43313cc --force --keep-logs 2026-03-09T17:53:46.324 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:53:47.678 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 01455850-1bdf-11f1-910a-9936d43313cc --force --keep-logs 2026-03-09T17:53:47.805 INFO:teuthology.orchestra.run.vm04.stdout:Deleting cluster with fsid: 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:53:48.897 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 01455850-1bdf-11f1-910a-9936d43313cc --force --keep-logs 2026-03-09T17:53:49.017 INFO:teuthology.orchestra.run.vm05.stdout:Deleting cluster with fsid: 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:53:50.333 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T17:53:50.360 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T17:53:50.391 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T17:53:50.430 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-09T17:53:50.431 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/590/remote/vm01/crash 2026-03-09T17:53:50.431 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/crash -- . 2026-03-09T17:53:50.454 INFO:teuthology.orchestra.run.vm01.stderr:tar: /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/crash: Cannot open: No such file or directory 2026-03-09T17:53:50.454 INFO:teuthology.orchestra.run.vm01.stderr:tar: Error is not recoverable: exiting now 2026-03-09T17:53:50.455 DEBUG:teuthology.misc:Transferring archived files from vm04:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/590/remote/vm04/crash 2026-03-09T17:53:50.455 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/crash -- . 2026-03-09T17:53:50.485 INFO:teuthology.orchestra.run.vm04.stderr:tar: /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/crash: Cannot open: No such file or directory 2026-03-09T17:53:50.485 INFO:teuthology.orchestra.run.vm04.stderr:tar: Error is not recoverable: exiting now 2026-03-09T17:53:50.486 DEBUG:teuthology.misc:Transferring archived files from vm05:/var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/590/remote/vm05/crash 2026-03-09T17:53:50.486 DEBUG:teuthology.orchestra.run.vm05:> sudo tar c -f - -C /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/crash -- . 2026-03-09T17:53:50.513 INFO:teuthology.orchestra.run.vm05.stderr:tar: /var/lib/ceph/01455850-1bdf-11f1-910a-9936d43313cc/crash: Cannot open: No such file or directory 2026-03-09T17:53:50.513 INFO:teuthology.orchestra.run.vm05.stderr:tar: Error is not recoverable: exiting now 2026-03-09T17:53:50.514 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-09T17:53:50.514 DEBUG:teuthology.orchestra.run.vm01:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/01455850-1bdf-11f1-910a-9936d43313cc/ceph.log | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v MON_DOWN | head -n 1 2026-03-09T17:53:50.537 INFO:teuthology.orchestra.run.vm01.stderr:grep: /var/log/ceph/01455850-1bdf-11f1-910a-9936d43313cc/ceph.log: No such file or directory 2026-03-09T17:53:50.539 WARNING:tasks.cephadm:Found errors (ERR|WRN|SEC) in cluster log 2026-03-09T17:53:50.539 DEBUG:teuthology.orchestra.run.vm01:> sudo egrep '\[SEC\]' /var/log/ceph/01455850-1bdf-11f1-910a-9936d43313cc/ceph.log | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v MON_DOWN | head -n 1 2026-03-09T17:53:50.602 INFO:teuthology.orchestra.run.vm01.stderr:grep: /var/log/ceph/01455850-1bdf-11f1-910a-9936d43313cc/ceph.log: No such file or directory 2026-03-09T17:53:50.603 INFO:tasks.cephadm:Compressing logs... 2026-03-09T17:53:50.604 DEBUG:teuthology.orchestra.run.vm01:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T17:53:50.646 DEBUG:teuthology.orchestra.run.vm04:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T17:53:50.647 DEBUG:teuthology.orchestra.run.vm05:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T17:53:50.668 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T17:53:50.668 INFO:teuthology.orchestra.run.vm01.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-09T17:53:50.670 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/cephadm.log: 87.2% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T17:53:50.671 INFO:teuthology.orchestra.run.vm04.stderr:find: ‘/var/log/rbd-target-api’gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T17:53:50.671 INFO:teuthology.orchestra.run.vm04.stderr:: No such file or directory 2026-03-09T17:53:50.671 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-09T17:53:50.671 INFO:teuthology.orchestra.run.vm01.stderr:real 0m0.013s 2026-03-09T17:53:50.671 INFO:teuthology.orchestra.run.vm01.stderr:user 0m0.007s 2026-03-09T17:53:50.671 INFO:teuthology.orchestra.run.vm01.stderr:sys 0m0.011s 2026-03-09T17:53:50.672 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/cephadm.log: 89.3% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T17:53:50.673 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T17:53:50.673 INFO:teuthology.orchestra.run.vm05.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-09T17:53:50.674 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-09T17:53:50.674 INFO:teuthology.orchestra.run.vm04.stderr:real 0m0.013s 2026-03-09T17:53:50.674 INFO:teuthology.orchestra.run.vm04.stderr:user 0m0.002s 2026-03-09T17:53:50.674 INFO:teuthology.orchestra.run.vm04.stderr:sys 0m0.018s 2026-03-09T17:53:50.676 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/cephadm.log: 90.1% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T17:53:50.677 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-09T17:53:50.677 INFO:teuthology.orchestra.run.vm05.stderr:real 0m0.015s 2026-03-09T17:53:50.677 INFO:teuthology.orchestra.run.vm05.stderr:user 0m0.006s 2026-03-09T17:53:50.677 INFO:teuthology.orchestra.run.vm05.stderr:sys 0m0.015s 2026-03-09T17:53:50.677 INFO:tasks.cephadm:Archiving logs... 2026-03-09T17:53:50.677 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/590/remote/vm01/log 2026-03-09T17:53:50.677 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T17:53:50.735 DEBUG:teuthology.misc:Transferring archived files from vm04:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/590/remote/vm04/log 2026-03-09T17:53:50.735 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T17:53:50.764 DEBUG:teuthology.misc:Transferring archived files from vm05:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/590/remote/vm05/log 2026-03-09T17:53:50.764 DEBUG:teuthology.orchestra.run.vm05:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T17:53:50.790 INFO:tasks.cephadm:Removing cluster... 2026-03-09T17:53:50.790 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 01455850-1bdf-11f1-910a-9936d43313cc --force 2026-03-09T17:53:50.915 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:53:51.006 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 01455850-1bdf-11f1-910a-9936d43313cc --force 2026-03-09T17:53:51.134 INFO:teuthology.orchestra.run.vm04.stdout:Deleting cluster with fsid: 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:53:51.226 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 01455850-1bdf-11f1-910a-9936d43313cc --force 2026-03-09T17:53:51.345 INFO:teuthology.orchestra.run.vm05.stdout:Deleting cluster with fsid: 01455850-1bdf-11f1-910a-9936d43313cc 2026-03-09T17:53:51.430 INFO:tasks.cephadm:Removing cephadm ... 2026-03-09T17:53:51.430 DEBUG:teuthology.orchestra.run.vm01:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T17:53:51.444 DEBUG:teuthology.orchestra.run.vm04:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T17:53:51.460 DEBUG:teuthology.orchestra.run.vm05:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T17:53:51.474 INFO:tasks.cephadm:Teardown complete 2026-03-09T17:53:51.475 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-09T17:53:51.477 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-09T17:53:51.477 DEBUG:teuthology.orchestra.run.vm01:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T17:53:51.486 DEBUG:teuthology.orchestra.run.vm04:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T17:53:51.498 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-09T17:53:51.503 DEBUG:teuthology.orchestra.run.vm05:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T17:53:51.515 INFO:teuthology.orchestra.run.vm04.stderr:bash: line 1: ntpq: command not found 2026-03-09T17:53:51.529 INFO:teuthology.orchestra.run.vm05.stderr:bash: line 1: ntpq: command not found 2026-03-09T17:53:51.786 INFO:teuthology.orchestra.run.vm04.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T17:53:51.786 INFO:teuthology.orchestra.run.vm04.stdout:=============================================================================== 2026-03-09T17:53:51.786 INFO:teuthology.orchestra.run.vm04.stdout:^+ ntp2.wup-de.hosts.301-mo> 2 6 377 238 -165us[ -391us] +/- 20ms 2026-03-09T17:53:51.786 INFO:teuthology.orchestra.run.vm04.stdout:^+ vps-fra8.orleans.ddnss.de 2 6 377 44 +122us[+6871ns] +/- 18ms 2026-03-09T17:53:51.786 INFO:teuthology.orchestra.run.vm04.stdout:^* vps-fra1.orleans.ddnss.de 2 6 377 44 -39us[ -154us] +/- 12ms 2026-03-09T17:53:51.786 INFO:teuthology.orchestra.run.vm04.stdout:^- y.ns.gin.ntt.net 2 6 377 45 +572us[ +456us] +/- 106ms 2026-03-09T17:53:51.786 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T17:53:51.786 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-09T17:53:51.787 INFO:teuthology.orchestra.run.vm01.stdout:^- y.ns.gin.ntt.net 2 6 377 44 +213us[ +213us] +/- 106ms 2026-03-09T17:53:51.787 INFO:teuthology.orchestra.run.vm01.stdout:^+ ntp2.wup-de.hosts.301-mo> 2 6 377 238 -552us[ -530us] +/- 20ms 2026-03-09T17:53:51.787 INFO:teuthology.orchestra.run.vm01.stdout:^* vps-fra1.orleans.ddnss.de 2 6 377 109 -473us[ -449us] +/- 12ms 2026-03-09T17:53:51.787 INFO:teuthology.orchestra.run.vm01.stdout:^+ vps-fra8.orleans.ddnss.de 2 6 377 45 +1107us[+1107us] +/- 16ms 2026-03-09T17:53:51.787 INFO:teuthology.orchestra.run.vm05.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T17:53:51.787 INFO:teuthology.orchestra.run.vm05.stdout:=============================================================================== 2026-03-09T17:53:51.787 INFO:teuthology.orchestra.run.vm05.stdout:^* vps-fra1.orleans.ddnss.de 2 6 377 45 -424us[ -463us] +/- 12ms 2026-03-09T17:53:51.787 INFO:teuthology.orchestra.run.vm05.stdout:^+ vps-fra8.orleans.ddnss.de 2 6 377 44 +175us[ +175us] +/- 18ms 2026-03-09T17:53:51.787 INFO:teuthology.orchestra.run.vm05.stdout:^- y.ns.gin.ntt.net 2 6 377 43 +240us[ +240us] +/- 106ms 2026-03-09T17:53:51.787 INFO:teuthology.orchestra.run.vm05.stdout:^+ ntp2.wup-de.hosts.301-mo> 2 6 377 42 -496us[ -496us] +/- 20ms 2026-03-09T17:53:51.787 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-09T17:53:51.790 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-09T17:53:51.790 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-09T17:53:51.792 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-09T17:53:51.795 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-09T17:53:51.798 INFO:teuthology.task.internal:Duration was 990.036891 seconds 2026-03-09T17:53:51.798 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-09T17:53:51.801 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-09T17:53:51.801 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T17:53:51.830 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T17:53:51.831 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T17:53:51.867 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T17:53:51.867 INFO:teuthology.orchestra.run.vm05.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T17:53:51.873 INFO:teuthology.orchestra.run.vm04.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T17:53:52.345 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-09T17:53:52.345 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm01.local 2026-03-09T17:53:52.345 DEBUG:teuthology.orchestra.run.vm01:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T17:53:52.369 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm04.local 2026-03-09T17:53:52.369 DEBUG:teuthology.orchestra.run.vm04:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T17:53:52.398 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm05.local 2026-03-09T17:53:52.398 DEBUG:teuthology.orchestra.run.vm05:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T17:53:52.421 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-09T17:53:52.421 DEBUG:teuthology.orchestra.run.vm01:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T17:53:52.424 DEBUG:teuthology.orchestra.run.vm04:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T17:53:52.441 DEBUG:teuthology.orchestra.run.vm05:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T17:53:52.954 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-09T17:53:52.954 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T17:53:52.956 DEBUG:teuthology.orchestra.run.vm04:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T17:53:52.957 DEBUG:teuthology.orchestra.run.vm05:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T17:53:52.977 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T17:53:52.977 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T17:53:52.978 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: gzip 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T17:53:52.978 INFO:teuthology.orchestra.run.vm01.stderr: -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T17:53:52.978 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-09T17:53:52.982 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T17:53:52.982 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T17:53:52.982 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T17:53:52.983 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T17:53:52.983 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: /home/ubuntu/cephtest/archive/syslog/journalctl.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-09T17:53:52.984 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T17:53:52.984 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T17:53:52.984 INFO:teuthology.orchestra.run.vm05.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T17:53:52.985 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T17:53:52.985 INFO:teuthology.orchestra.run.vm05.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: /home/ubuntu/cephtest/archive/syslog/journalctl.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-09T17:53:53.127 INFO:teuthology.orchestra.run.vm04.stderr: 98.4% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T17:53:53.134 INFO:teuthology.orchestra.run.vm05.stderr: 98.3% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T17:53:53.152 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.2% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T17:53:53.154 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-09T17:53:53.157 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-09T17:53:53.157 DEBUG:teuthology.orchestra.run.vm01:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T17:53:53.219 DEBUG:teuthology.orchestra.run.vm04:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T17:53:53.245 DEBUG:teuthology.orchestra.run.vm05:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T17:53:53.268 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-09T17:53:53.271 DEBUG:teuthology.orchestra.run.vm01:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T17:53:53.273 DEBUG:teuthology.orchestra.run.vm04:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T17:53:53.287 DEBUG:teuthology.orchestra.run.vm05:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T17:53:53.299 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = core 2026-03-09T17:53:53.312 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern = core 2026-03-09T17:53:53.333 INFO:teuthology.orchestra.run.vm05.stdout:kernel.core_pattern = core 2026-03-09T17:53:53.347 DEBUG:teuthology.orchestra.run.vm01:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T17:53:53.366 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:53:53.366 DEBUG:teuthology.orchestra.run.vm04:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T17:53:53.382 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:53:53.382 DEBUG:teuthology.orchestra.run.vm05:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T17:53:53.403 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:53:53.403 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-09T17:53:53.405 INFO:teuthology.task.internal:Transferring archived files... 2026-03-09T17:53:53.406 DEBUG:teuthology.misc:Transferring archived files from vm01:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/590/remote/vm01 2026-03-09T17:53:53.406 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T17:53:53.437 DEBUG:teuthology.misc:Transferring archived files from vm04:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/590/remote/vm04 2026-03-09T17:53:53.438 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T17:53:53.468 DEBUG:teuthology.misc:Transferring archived files from vm05:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/590/remote/vm05 2026-03-09T17:53:53.468 DEBUG:teuthology.orchestra.run.vm05:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T17:53:53.495 INFO:teuthology.task.internal:Removing archive directory... 2026-03-09T17:53:53.495 DEBUG:teuthology.orchestra.run.vm01:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T17:53:53.497 DEBUG:teuthology.orchestra.run.vm04:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T17:53:53.510 DEBUG:teuthology.orchestra.run.vm05:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T17:53:53.549 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-09T17:53:53.552 INFO:teuthology.task.internal:Not uploading archives. 2026-03-09T17:53:53.552 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-09T17:53:53.555 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-09T17:53:53.555 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T17:53:53.556 DEBUG:teuthology.orchestra.run.vm04:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T17:53:53.566 DEBUG:teuthology.orchestra.run.vm05:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T17:53:53.572 INFO:teuthology.orchestra.run.vm01.stdout: 8532143 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 17:53 /home/ubuntu/cephtest 2026-03-09T17:53:53.581 INFO:teuthology.orchestra.run.vm04.stdout: 8532144 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 17:53 /home/ubuntu/cephtest 2026-03-09T17:53:53.608 INFO:teuthology.orchestra.run.vm05.stdout: 8532143 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 17:53 /home/ubuntu/cephtest 2026-03-09T17:53:53.609 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-09T17:53:53.615 INFO:teuthology.run:Summary data: description: orch/cephadm/rbd_iscsi/{base/install cluster/{fixed-3 openstack} conf/{disable-pool-app} supported-container-hosts$/{centos_9.stream} workloads/cephadm_iscsi} duration: 990.0368914604187 failure_reason: '"grep: /var/log/ceph/01455850-1bdf-11f1-910a-9936d43313cc/ceph.log: No such file or directory" in cluster log' flavor: default owner: kyr success: false 2026-03-09T17:53:53.615 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-09T17:53:53.638 INFO:teuthology.run:FAIL