2026-03-10T05:25:40.587 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-10T05:25:40.591 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T05:25:40.607 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/914 branch: squid description: orch/cephadm/rbd_iscsi/{base/install cluster/{fixed-3 openstack} conf/{disable-pool-app} supported-container-hosts$/{centos_9.stream_runc} workloads/cephadm_iscsi} email: null first_in_suite: false flavor: default job_id: '914' last_in_suite: false machine_type: vps name: kyr-2026-03-10_01:00:38-orch-squid-none-default-vps no_nested_subset: false openstack: - machine: cpus: 1 disk: 40 ram: 8000 volumes: count: 4 size: 30 os_type: centos os_version: 9.stream overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: global: mon warn on pool no app: false mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - MON_DOWN sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 workunit: branch: tt-squid sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - mon.a - mgr.x - osd.0 - osd.1 - client.0 - ceph.iscsi.iscsi.a - - mon.b - osd.2 - osd.3 - osd.4 - client.1 - - mon.c - osd.5 - osd.6 - osd.7 - client.2 - ceph.iscsi.iscsi.b seed: 8043 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b targets: vm01.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOj68T01R/h9YAAwK2VZ+SB2kJVHjxcqkxmJe4biW5xuoXqWAM/iVrZVdj5LaHuxgGZWjuBjJdPtVf8rmsy6fGI= vm02.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLWnKN4cpUejhtCfv4P4m2fvmg8TcJIwa3L1fpFK8rhAims2SsuGap0bl4AKIMbdgbzfhRNQMt2FVENQy/s8F2k= vm05.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLLB5jyWEGsrlmWIKB5hZs6emuaOPj3bnQkxXIjIxBrA1GarL6+QoCUXGYgQiOBu1qorUcXFCZ0jo68Ycldvb08= tasks: - cephadm: null - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - install: extra_system_packages: deb: - open-iscsi - multipath-tools rpm: - iscsi-initiator-utils - device-mapper-multipath - pexec: all: - sudo dnf remove nvme-cli -y - sudo dnf install runc nvmetcli nvme-cli -y - sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf - sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf - ceph_iscsi_client: clients: - client.1 - cram: clients: client.0: - src/test/cli-integration/rbd/gwcli_create.t client.1: - src/test/cli-integration/rbd/iscsi_client.t client.2: - src/test/cli-integration/rbd/gwcli_delete.t parallel: false - cram: clients: client.0: - src/test/cli-integration/rbd/rest_api_create.t client.1: - src/test/cli-integration/rbd/iscsi_client.t client.2: - src/test/cli-integration/rbd/rest_api_delete.t parallel: false teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-10_01:00:38 tube: vps use_shaman: true user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.611473 2026-03-10T05:25:40.608 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa; will attempt to use it 2026-03-10T05:25:40.608 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa/tasks 2026-03-10T05:25:40.608 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-10T05:25:40.608 INFO:teuthology.task.internal:Checking packages... 2026-03-10T05:25:40.608 INFO:teuthology.task.internal:Checking packages for os_type 'centos', flavor 'default' and ceph hash 'e911bdebe5c8faa3800735d1568fcdca65db60df' 2026-03-10T05:25:40.608 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-10T05:25:40.608 INFO:teuthology.packaging:ref: None 2026-03-10T05:25:40.608 INFO:teuthology.packaging:tag: None 2026-03-10T05:25:40.608 INFO:teuthology.packaging:branch: squid 2026-03-10T05:25:40.608 INFO:teuthology.packaging:sha1: e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:25:40.608 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&ref=squid 2026-03-10T05:25:41.416 INFO:teuthology.task.internal:Found packages for ceph version 19.2.3-678.ge911bdeb 2026-03-10T05:25:41.417 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-10T05:25:41.417 INFO:teuthology.task.internal:no buildpackages task found 2026-03-10T05:25:41.417 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-10T05:25:41.418 INFO:teuthology.task.internal:Saving configuration 2026-03-10T05:25:41.422 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-10T05:25:41.423 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-10T05:25:41.430 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm01.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/914', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 05:24:07.692931', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:01', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOj68T01R/h9YAAwK2VZ+SB2kJVHjxcqkxmJe4biW5xuoXqWAM/iVrZVdj5LaHuxgGZWjuBjJdPtVf8rmsy6fGI='} 2026-03-10T05:25:41.435 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm02.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/914', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 05:24:07.693759', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:02', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLWnKN4cpUejhtCfv4P4m2fvmg8TcJIwa3L1fpFK8rhAims2SsuGap0bl4AKIMbdgbzfhRNQMt2FVENQy/s8F2k='} 2026-03-10T05:25:41.439 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm05.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/914', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 05:24:07.693981', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:05', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLLB5jyWEGsrlmWIKB5hZs6emuaOPj3bnQkxXIjIxBrA1GarL6+QoCUXGYgQiOBu1qorUcXFCZ0jo68Ycldvb08='} 2026-03-10T05:25:41.439 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-10T05:25:41.440 INFO:teuthology.task.internal:roles: ubuntu@vm01.local - ['host.a', 'mon.a', 'mgr.x', 'osd.0', 'osd.1', 'client.0', 'ceph.iscsi.iscsi.a'] 2026-03-10T05:25:41.440 INFO:teuthology.task.internal:roles: ubuntu@vm02.local - ['mon.b', 'osd.2', 'osd.3', 'osd.4', 'client.1'] 2026-03-10T05:25:41.440 INFO:teuthology.task.internal:roles: ubuntu@vm05.local - ['mon.c', 'osd.5', 'osd.6', 'osd.7', 'client.2', 'ceph.iscsi.iscsi.b'] 2026-03-10T05:25:41.440 INFO:teuthology.run_tasks:Running task console_log... 2026-03-10T05:25:41.445 DEBUG:teuthology.task.console_log:vm01 does not support IPMI; excluding 2026-03-10T05:25:41.450 DEBUG:teuthology.task.console_log:vm02 does not support IPMI; excluding 2026-03-10T05:25:41.454 DEBUG:teuthology.task.console_log:vm05 does not support IPMI; excluding 2026-03-10T05:25:41.455 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7ff65eae3e20>, signals=[15]) 2026-03-10T05:25:41.455 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-10T05:25:41.455 INFO:teuthology.task.internal:Opening connections... 2026-03-10T05:25:41.455 DEBUG:teuthology.task.internal:connecting to ubuntu@vm01.local 2026-03-10T05:25:41.456 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T05:25:41.515 DEBUG:teuthology.task.internal:connecting to ubuntu@vm02.local 2026-03-10T05:25:41.516 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T05:25:41.575 DEBUG:teuthology.task.internal:connecting to ubuntu@vm05.local 2026-03-10T05:25:41.575 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm05.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T05:25:41.632 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-10T05:25:41.633 DEBUG:teuthology.orchestra.run.vm01:> uname -m 2026-03-10T05:25:41.647 INFO:teuthology.orchestra.run.vm01.stdout:x86_64 2026-03-10T05:25:41.647 DEBUG:teuthology.orchestra.run.vm01:> cat /etc/os-release 2026-03-10T05:25:41.703 INFO:teuthology.orchestra.run.vm01.stdout:NAME="CentOS Stream" 2026-03-10T05:25:41.703 INFO:teuthology.orchestra.run.vm01.stdout:VERSION="9" 2026-03-10T05:25:41.703 INFO:teuthology.orchestra.run.vm01.stdout:ID="centos" 2026-03-10T05:25:41.703 INFO:teuthology.orchestra.run.vm01.stdout:ID_LIKE="rhel fedora" 2026-03-10T05:25:41.703 INFO:teuthology.orchestra.run.vm01.stdout:VERSION_ID="9" 2026-03-10T05:25:41.703 INFO:teuthology.orchestra.run.vm01.stdout:PLATFORM_ID="platform:el9" 2026-03-10T05:25:41.703 INFO:teuthology.orchestra.run.vm01.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-10T05:25:41.703 INFO:teuthology.orchestra.run.vm01.stdout:ANSI_COLOR="0;31" 2026-03-10T05:25:41.703 INFO:teuthology.orchestra.run.vm01.stdout:LOGO="fedora-logo-icon" 2026-03-10T05:25:41.703 INFO:teuthology.orchestra.run.vm01.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-10T05:25:41.703 INFO:teuthology.orchestra.run.vm01.stdout:HOME_URL="https://centos.org/" 2026-03-10T05:25:41.703 INFO:teuthology.orchestra.run.vm01.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-10T05:25:41.703 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-10T05:25:41.703 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-10T05:25:41.703 INFO:teuthology.lock.ops:Updating vm01.local on lock server 2026-03-10T05:25:41.708 DEBUG:teuthology.orchestra.run.vm02:> uname -m 2026-03-10T05:25:41.722 INFO:teuthology.orchestra.run.vm02.stdout:x86_64 2026-03-10T05:25:41.722 DEBUG:teuthology.orchestra.run.vm02:> cat /etc/os-release 2026-03-10T05:25:41.776 INFO:teuthology.orchestra.run.vm02.stdout:NAME="CentOS Stream" 2026-03-10T05:25:41.776 INFO:teuthology.orchestra.run.vm02.stdout:VERSION="9" 2026-03-10T05:25:41.776 INFO:teuthology.orchestra.run.vm02.stdout:ID="centos" 2026-03-10T05:25:41.776 INFO:teuthology.orchestra.run.vm02.stdout:ID_LIKE="rhel fedora" 2026-03-10T05:25:41.776 INFO:teuthology.orchestra.run.vm02.stdout:VERSION_ID="9" 2026-03-10T05:25:41.776 INFO:teuthology.orchestra.run.vm02.stdout:PLATFORM_ID="platform:el9" 2026-03-10T05:25:41.776 INFO:teuthology.orchestra.run.vm02.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-10T05:25:41.776 INFO:teuthology.orchestra.run.vm02.stdout:ANSI_COLOR="0;31" 2026-03-10T05:25:41.776 INFO:teuthology.orchestra.run.vm02.stdout:LOGO="fedora-logo-icon" 2026-03-10T05:25:41.776 INFO:teuthology.orchestra.run.vm02.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-10T05:25:41.776 INFO:teuthology.orchestra.run.vm02.stdout:HOME_URL="https://centos.org/" 2026-03-10T05:25:41.776 INFO:teuthology.orchestra.run.vm02.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-10T05:25:41.776 INFO:teuthology.orchestra.run.vm02.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-10T05:25:41.776 INFO:teuthology.orchestra.run.vm02.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-10T05:25:41.776 INFO:teuthology.lock.ops:Updating vm02.local on lock server 2026-03-10T05:25:41.780 DEBUG:teuthology.orchestra.run.vm05:> uname -m 2026-03-10T05:25:41.794 INFO:teuthology.orchestra.run.vm05.stdout:x86_64 2026-03-10T05:25:41.794 DEBUG:teuthology.orchestra.run.vm05:> cat /etc/os-release 2026-03-10T05:25:41.848 INFO:teuthology.orchestra.run.vm05.stdout:NAME="CentOS Stream" 2026-03-10T05:25:41.848 INFO:teuthology.orchestra.run.vm05.stdout:VERSION="9" 2026-03-10T05:25:41.848 INFO:teuthology.orchestra.run.vm05.stdout:ID="centos" 2026-03-10T05:25:41.848 INFO:teuthology.orchestra.run.vm05.stdout:ID_LIKE="rhel fedora" 2026-03-10T05:25:41.848 INFO:teuthology.orchestra.run.vm05.stdout:VERSION_ID="9" 2026-03-10T05:25:41.848 INFO:teuthology.orchestra.run.vm05.stdout:PLATFORM_ID="platform:el9" 2026-03-10T05:25:41.848 INFO:teuthology.orchestra.run.vm05.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-10T05:25:41.848 INFO:teuthology.orchestra.run.vm05.stdout:ANSI_COLOR="0;31" 2026-03-10T05:25:41.848 INFO:teuthology.orchestra.run.vm05.stdout:LOGO="fedora-logo-icon" 2026-03-10T05:25:41.848 INFO:teuthology.orchestra.run.vm05.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-10T05:25:41.848 INFO:teuthology.orchestra.run.vm05.stdout:HOME_URL="https://centos.org/" 2026-03-10T05:25:41.848 INFO:teuthology.orchestra.run.vm05.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-10T05:25:41.848 INFO:teuthology.orchestra.run.vm05.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-10T05:25:41.848 INFO:teuthology.orchestra.run.vm05.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-10T05:25:41.848 INFO:teuthology.lock.ops:Updating vm05.local on lock server 2026-03-10T05:25:41.852 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-10T05:25:41.854 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-10T05:25:41.855 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-10T05:25:41.855 DEBUG:teuthology.orchestra.run.vm01:> test '!' -e /home/ubuntu/cephtest 2026-03-10T05:25:41.856 DEBUG:teuthology.orchestra.run.vm02:> test '!' -e /home/ubuntu/cephtest 2026-03-10T05:25:41.858 DEBUG:teuthology.orchestra.run.vm05:> test '!' -e /home/ubuntu/cephtest 2026-03-10T05:25:41.902 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-10T05:25:41.903 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-10T05:25:41.903 DEBUG:teuthology.orchestra.run.vm01:> test -z $(ls -A /var/lib/ceph) 2026-03-10T05:25:41.911 DEBUG:teuthology.orchestra.run.vm02:> test -z $(ls -A /var/lib/ceph) 2026-03-10T05:25:41.912 DEBUG:teuthology.orchestra.run.vm05:> test -z $(ls -A /var/lib/ceph) 2026-03-10T05:25:41.924 INFO:teuthology.orchestra.run.vm01.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T05:25:41.927 INFO:teuthology.orchestra.run.vm02.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T05:25:41.957 INFO:teuthology.orchestra.run.vm05.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T05:25:41.957 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-10T05:25:41.965 DEBUG:teuthology.orchestra.run.vm01:> test -e /ceph-qa-ready 2026-03-10T05:25:41.978 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:25:42.165 DEBUG:teuthology.orchestra.run.vm02:> test -e /ceph-qa-ready 2026-03-10T05:25:42.179 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:25:42.370 DEBUG:teuthology.orchestra.run.vm05:> test -e /ceph-qa-ready 2026-03-10T05:25:42.383 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:25:42.597 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-10T05:25:42.599 INFO:teuthology.task.internal:Creating test directory... 2026-03-10T05:25:42.599 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T05:25:42.601 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T05:25:42.603 DEBUG:teuthology.orchestra.run.vm05:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T05:25:42.618 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-10T05:25:42.620 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-10T05:25:42.621 INFO:teuthology.task.internal:Creating archive directory... 2026-03-10T05:25:42.621 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T05:25:42.657 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T05:25:42.660 DEBUG:teuthology.orchestra.run.vm05:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T05:25:42.680 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-10T05:25:42.681 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-10T05:25:42.681 DEBUG:teuthology.orchestra.run.vm01:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T05:25:42.725 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:25:42.726 DEBUG:teuthology.orchestra.run.vm02:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T05:25:42.742 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:25:42.742 DEBUG:teuthology.orchestra.run.vm05:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T05:25:42.756 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:25:42.756 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T05:25:42.768 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T05:25:42.783 DEBUG:teuthology.orchestra.run.vm05:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T05:25:42.789 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T05:25:42.798 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T05:25:42.807 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T05:25:42.815 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T05:25:42.820 INFO:teuthology.orchestra.run.vm05.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T05:25:42.828 INFO:teuthology.orchestra.run.vm05.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T05:25:42.829 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-10T05:25:42.830 INFO:teuthology.task.internal:Configuring sudo... 2026-03-10T05:25:42.831 DEBUG:teuthology.orchestra.run.vm01:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T05:25:42.841 DEBUG:teuthology.orchestra.run.vm02:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T05:25:42.858 DEBUG:teuthology.orchestra.run.vm05:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T05:25:42.893 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-10T05:25:42.895 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-10T05:25:42.895 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T05:25:42.908 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T05:25:42.921 DEBUG:teuthology.orchestra.run.vm05:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T05:25:42.947 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T05:25:42.981 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T05:25:43.040 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:25:43.040 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T05:25:43.093 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T05:25:43.115 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T05:25:43.171 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:25:43.171 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T05:25:43.229 DEBUG:teuthology.orchestra.run.vm05:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T05:25:43.251 DEBUG:teuthology.orchestra.run.vm05:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T05:25:43.307 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T05:25:43.307 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T05:25:43.364 DEBUG:teuthology.orchestra.run.vm01:> sudo service rsyslog restart 2026-03-10T05:25:43.365 DEBUG:teuthology.orchestra.run.vm02:> sudo service rsyslog restart 2026-03-10T05:25:43.367 DEBUG:teuthology.orchestra.run.vm05:> sudo service rsyslog restart 2026-03-10T05:25:43.390 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T05:25:43.391 INFO:teuthology.orchestra.run.vm02.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T05:25:43.430 INFO:teuthology.orchestra.run.vm05.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T05:25:43.883 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-10T05:25:43.885 INFO:teuthology.task.internal:Starting timer... 2026-03-10T05:25:43.885 INFO:teuthology.run_tasks:Running task pcp... 2026-03-10T05:25:43.888 INFO:teuthology.run_tasks:Running task selinux... 2026-03-10T05:25:43.890 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0']} 2026-03-10T05:25:43.890 INFO:teuthology.task.selinux:Excluding vm01: VMs are not yet supported 2026-03-10T05:25:43.890 INFO:teuthology.task.selinux:Excluding vm02: VMs are not yet supported 2026-03-10T05:25:43.890 INFO:teuthology.task.selinux:Excluding vm05: VMs are not yet supported 2026-03-10T05:25:43.890 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-10T05:25:43.890 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-10T05:25:43.890 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-10T05:25:43.890 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-10T05:25:43.891 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-10T05:25:43.892 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-10T05:25:43.893 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-10T05:25:44.501 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-10T05:25:44.506 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-10T05:25:44.506 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventoryzsfr51ro --limit vm01.local,vm02.local,vm05.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-10T05:27:40.286 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm01.local'), Remote(name='ubuntu@vm02.local'), Remote(name='ubuntu@vm05.local')] 2026-03-10T05:27:40.286 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm01.local' 2026-03-10T05:27:40.287 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T05:27:40.352 DEBUG:teuthology.orchestra.run.vm01:> true 2026-03-10T05:27:40.438 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm01.local' 2026-03-10T05:27:40.438 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm02.local' 2026-03-10T05:27:40.438 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T05:27:40.502 DEBUG:teuthology.orchestra.run.vm02:> true 2026-03-10T05:27:40.581 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm02.local' 2026-03-10T05:27:40.581 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm05.local' 2026-03-10T05:27:40.581 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm05.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T05:27:40.649 DEBUG:teuthology.orchestra.run.vm05:> true 2026-03-10T05:27:40.727 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm05.local' 2026-03-10T05:27:40.727 INFO:teuthology.run_tasks:Running task clock... 2026-03-10T05:27:40.729 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-10T05:27:40.729 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T05:27:40.729 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T05:27:40.732 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T05:27:40.733 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T05:27:40.734 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T05:27:40.734 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T05:27:40.768 INFO:teuthology.orchestra.run.vm02.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-10T05:27:40.769 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-10T05:27:40.780 INFO:teuthology.orchestra.run.vm02.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-10T05:27:40.785 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-10T05:27:40.804 INFO:teuthology.orchestra.run.vm02.stderr:sudo: ntpd: command not found 2026-03-10T05:27:40.809 INFO:teuthology.orchestra.run.vm01.stderr:sudo: ntpd: command not found 2026-03-10T05:27:40.817 INFO:teuthology.orchestra.run.vm05.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-10T05:27:40.817 INFO:teuthology.orchestra.run.vm02.stdout:506 Cannot talk to daemon 2026-03-10T05:27:40.823 INFO:teuthology.orchestra.run.vm01.stdout:506 Cannot talk to daemon 2026-03-10T05:27:40.831 INFO:teuthology.orchestra.run.vm02.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-10T05:27:40.842 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-10T05:27:40.845 INFO:teuthology.orchestra.run.vm05.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-10T05:27:40.851 INFO:teuthology.orchestra.run.vm02.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-10T05:27:40.857 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-10T05:27:40.882 INFO:teuthology.orchestra.run.vm05.stderr:sudo: ntpd: command not found 2026-03-10T05:27:40.897 INFO:teuthology.orchestra.run.vm02.stderr:bash: line 1: ntpq: command not found 2026-03-10T05:27:40.898 INFO:teuthology.orchestra.run.vm05.stdout:506 Cannot talk to daemon 2026-03-10T05:27:40.904 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-10T05:27:40.919 INFO:teuthology.orchestra.run.vm05.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-10T05:27:40.944 INFO:teuthology.orchestra.run.vm05.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-10T05:27:40.947 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T05:27:40.947 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-10T05:27:40.947 INFO:teuthology.orchestra.run.vm01.stdout:^? www.festibal.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:27:40.947 INFO:teuthology.orchestra.run.vm01.stdout:^? mail.anyvm.tech 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:27:40.947 INFO:teuthology.orchestra.run.vm01.stdout:^? frank.askja.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:27:40.947 INFO:teuthology.orchestra.run.vm01.stdout:^? 139-162-156-95.ip.linode> 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:27:40.947 INFO:teuthology.orchestra.run.vm02.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T05:27:40.947 INFO:teuthology.orchestra.run.vm02.stdout:=============================================================================== 2026-03-10T05:27:40.947 INFO:teuthology.orchestra.run.vm02.stdout:^? 139-162-156-95.ip.linode> 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:27:40.947 INFO:teuthology.orchestra.run.vm02.stdout:^? www.festibal.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:27:40.947 INFO:teuthology.orchestra.run.vm02.stdout:^? mail.anyvm.tech 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:27:40.947 INFO:teuthology.orchestra.run.vm02.stdout:^? frank.askja.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:27:41.001 INFO:teuthology.orchestra.run.vm05.stderr:bash: line 1: ntpq: command not found 2026-03-10T05:27:41.004 INFO:teuthology.orchestra.run.vm05.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T05:27:41.004 INFO:teuthology.orchestra.run.vm05.stdout:=============================================================================== 2026-03-10T05:27:41.004 INFO:teuthology.orchestra.run.vm05.stdout:^? 139-162-156-95.ip.linode> 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:27:41.004 INFO:teuthology.orchestra.run.vm05.stdout:^? www.festibal.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:27:41.004 INFO:teuthology.orchestra.run.vm05.stdout:^? mail.anyvm.tech 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:27:41.004 INFO:teuthology.orchestra.run.vm05.stdout:^? frank.askja.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T05:27:41.004 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-10T05:27:41.050 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon warn on pool no app': False}, 'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'MON_DOWN'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-10T05:27:41.050 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:27:41.050 INFO:tasks.cephadm:Cluster fsid is dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:27:41.050 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-10T05:27:41.050 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.101', 'mon.b': '192.168.123.102', 'mon.c': '192.168.123.105'} 2026-03-10T05:27:41.050 INFO:tasks.cephadm:First mon is mon.a on vm01 2026-03-10T05:27:41.050 INFO:tasks.cephadm:First mgr is x 2026-03-10T05:27:41.050 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-10T05:27:41.050 DEBUG:teuthology.orchestra.run.vm01:> sudo hostname $(hostname -s) 2026-03-10T05:27:41.089 DEBUG:teuthology.orchestra.run.vm02:> sudo hostname $(hostname -s) 2026-03-10T05:27:41.127 DEBUG:teuthology.orchestra.run.vm05:> sudo hostname $(hostname -s) 2026-03-10T05:27:41.161 INFO:tasks.cephadm:Downloading "compiled" cephadm from cachra 2026-03-10T05:27:41.162 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:27:41.747 INFO:tasks.cephadm:builder_project result: [{'url': 'https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'chacra_url': 'https://3.chacra.ceph.com/repos/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'ref': 'squid', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'distro': 'centos', 'distro_version': '9', 'distro_codename': None, 'modified': '2026-02-25 18:55:15.146628', 'status': 'ready', 'flavor': 'default', 'project': 'ceph', 'archs': ['source', 'x86_64'], 'extra': {'version': '19.2.3-678-ge911bdeb', 'package_manager_version': '19.2.3-678.ge911bdeb', 'build_url': 'https://jenkins.ceph.com/job/ceph-dev-pipeline/3275/', 'root_build_cause': '', 'node_name': '10.20.192.26+soko16', 'job_name': 'ceph-dev-pipeline'}}] 2026-03-10T05:27:42.394 INFO:tasks.util.chacra:got chacra host 3.chacra.ceph.com, ref squid, sha1 e911bdebe5c8faa3800735d1568fcdca65db60df from https://shaman.ceph.com/api/search/?project=ceph&distros=centos%2F9%2Fx86_64&flavor=default&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:27:42.395 INFO:tasks.cephadm:Discovered cachra url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-10T05:27:42.395 INFO:tasks.cephadm:Downloading cephadm from url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-10T05:27:42.395 DEBUG:teuthology.orchestra.run.vm01:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T05:27:43.907 INFO:teuthology.orchestra.run.vm01.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 10 05:27 /home/ubuntu/cephtest/cephadm 2026-03-10T05:27:43.907 DEBUG:teuthology.orchestra.run.vm02:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T05:27:45.265 INFO:teuthology.orchestra.run.vm02.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 10 05:27 /home/ubuntu/cephtest/cephadm 2026-03-10T05:27:45.265 DEBUG:teuthology.orchestra.run.vm05:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T05:27:46.603 INFO:teuthology.orchestra.run.vm05.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 10 05:27 /home/ubuntu/cephtest/cephadm 2026-03-10T05:27:46.603 DEBUG:teuthology.orchestra.run.vm01:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T05:27:46.628 DEBUG:teuthology.orchestra.run.vm02:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T05:27:46.654 DEBUG:teuthology.orchestra.run.vm05:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T05:27:46.680 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df on all hosts... 2026-03-10T05:27:46.680 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T05:27:46.681 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T05:27:46.698 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T05:27:46.923 INFO:teuthology.orchestra.run.vm02.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T05:27:46.924 INFO:teuthology.orchestra.run.vm01.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T05:27:46.974 INFO:teuthology.orchestra.run.vm05.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T05:28:39.461 INFO:teuthology.orchestra.run.vm02.stdout:{ 2026-03-10T05:28:39.461 INFO:teuthology.orchestra.run.vm02.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T05:28:39.461 INFO:teuthology.orchestra.run.vm02.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T05:28:39.461 INFO:teuthology.orchestra.run.vm02.stdout: "repo_digests": [ 2026-03-10T05:28:39.461 INFO:teuthology.orchestra.run.vm02.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T05:28:39.461 INFO:teuthology.orchestra.run.vm02.stdout: ] 2026-03-10T05:28:39.461 INFO:teuthology.orchestra.run.vm02.stdout:} 2026-03-10T05:28:40.334 INFO:teuthology.orchestra.run.vm01.stdout:{ 2026-03-10T05:28:40.334 INFO:teuthology.orchestra.run.vm01.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T05:28:40.334 INFO:teuthology.orchestra.run.vm01.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T05:28:40.334 INFO:teuthology.orchestra.run.vm01.stdout: "repo_digests": [ 2026-03-10T05:28:40.334 INFO:teuthology.orchestra.run.vm01.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T05:28:40.334 INFO:teuthology.orchestra.run.vm01.stdout: ] 2026-03-10T05:28:40.334 INFO:teuthology.orchestra.run.vm01.stdout:} 2026-03-10T05:28:44.628 INFO:teuthology.orchestra.run.vm05.stdout:{ 2026-03-10T05:28:44.628 INFO:teuthology.orchestra.run.vm05.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T05:28:44.628 INFO:teuthology.orchestra.run.vm05.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T05:28:44.628 INFO:teuthology.orchestra.run.vm05.stdout: "repo_digests": [ 2026-03-10T05:28:44.628 INFO:teuthology.orchestra.run.vm05.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T05:28:44.628 INFO:teuthology.orchestra.run.vm05.stdout: ] 2026-03-10T05:28:44.628 INFO:teuthology.orchestra.run.vm05.stdout:} 2026-03-10T05:28:44.661 DEBUG:teuthology.orchestra.run.vm01:> sudo mkdir -p /etc/ceph 2026-03-10T05:28:44.689 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /etc/ceph 2026-03-10T05:28:44.729 DEBUG:teuthology.orchestra.run.vm05:> sudo mkdir -p /etc/ceph 2026-03-10T05:28:44.768 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod 777 /etc/ceph 2026-03-10T05:28:44.797 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 777 /etc/ceph 2026-03-10T05:28:44.830 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod 777 /etc/ceph 2026-03-10T05:28:44.867 INFO:tasks.cephadm:Writing seed config... 2026-03-10T05:28:44.867 INFO:tasks.cephadm: override: [global] mon warn on pool no app = False 2026-03-10T05:28:44.867 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-10T05:28:44.867 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-10T05:28:44.867 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-10T05:28:44.867 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-10T05:28:44.867 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-10T05:28:44.867 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-10T05:28:44.867 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-10T05:28:44.867 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-10T05:28:44.867 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:28:44.867 DEBUG:teuthology.orchestra.run.vm01:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-10T05:28:44.889 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba mon warn on pool no app = False [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-10T05:28:44.889 DEBUG:teuthology.orchestra.run.vm01:mon.a> sudo journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.a.service 2026-03-10T05:28:44.931 DEBUG:teuthology.orchestra.run.vm01:mgr.x> sudo journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mgr.x.service 2026-03-10T05:28:44.973 INFO:tasks.cephadm:Bootstrapping... 2026-03-10T05:28:44.973 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df -v bootstrap --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id x --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.101 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:28:45.128 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-10T05:28:45.128 INFO:teuthology.orchestra.run.vm01.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df', '-v', 'bootstrap', '--fsid', 'dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'x', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.101', '--skip-admin-label'] 2026-03-10T05:28:45.129 INFO:teuthology.orchestra.run.vm01.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-10T05:28:45.129 INFO:teuthology.orchestra.run.vm01.stdout:Verifying podman|docker is present... 2026-03-10T05:28:45.152 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 5.8.0 2026-03-10T05:28:45.152 INFO:teuthology.orchestra.run.vm01.stdout:Verifying lvm2 is present... 2026-03-10T05:28:45.152 INFO:teuthology.orchestra.run.vm01.stdout:Verifying time synchronization is in place... 2026-03-10T05:28:45.160 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-10T05:28:45.160 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T05:28:45.167 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-10T05:28:45.167 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-10T05:28:45.174 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout enabled 2026-03-10T05:28:45.180 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout active 2026-03-10T05:28:45.180 INFO:teuthology.orchestra.run.vm01.stdout:Unit chronyd.service is enabled and running 2026-03-10T05:28:45.180 INFO:teuthology.orchestra.run.vm01.stdout:Repeating the final host check... 2026-03-10T05:28:45.203 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 5.8.0 2026-03-10T05:28:45.203 INFO:teuthology.orchestra.run.vm01.stdout:podman (/bin/podman) version 5.8.0 is present 2026-03-10T05:28:45.204 INFO:teuthology.orchestra.run.vm01.stdout:systemctl is present 2026-03-10T05:28:45.204 INFO:teuthology.orchestra.run.vm01.stdout:lvcreate is present 2026-03-10T05:28:45.213 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-10T05:28:45.213 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T05:28:45.220 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-10T05:28:45.220 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout inactive 2026-03-10T05:28:45.227 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout enabled 2026-03-10T05:28:45.233 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stdout active 2026-03-10T05:28:45.234 INFO:teuthology.orchestra.run.vm01.stdout:Unit chronyd.service is enabled and running 2026-03-10T05:28:45.234 INFO:teuthology.orchestra.run.vm01.stdout:Host looks OK 2026-03-10T05:28:45.234 INFO:teuthology.orchestra.run.vm01.stdout:Cluster fsid: dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:28:45.234 INFO:teuthology.orchestra.run.vm01.stdout:Acquiring lock 140647385627088 on /run/cephadm/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba.lock 2026-03-10T05:28:45.234 INFO:teuthology.orchestra.run.vm01.stdout:Lock 140647385627088 acquired on /run/cephadm/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba.lock 2026-03-10T05:28:45.234 INFO:teuthology.orchestra.run.vm01.stdout:Verifying IP 192.168.123.101 port 3300 ... 2026-03-10T05:28:45.235 INFO:teuthology.orchestra.run.vm01.stdout:Verifying IP 192.168.123.101 port 6789 ... 2026-03-10T05:28:45.235 INFO:teuthology.orchestra.run.vm01.stdout:Base mon IP(s) is [192.168.123.101:3300, 192.168.123.101:6789], mon addrv is [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-10T05:28:45.239 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.101 metric 100 2026-03-10T05:28:45.239 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.101 metric 100 2026-03-10T05:28:45.242 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-10T05:28:45.243 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-10T05:28:45.246 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-10T05:28:45.247 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-10T05:28:45.247 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-10T05:28:45.247 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout 2: eth0: mtu 1500 state UP qlen 1000 2026-03-10T05:28:45.247 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout inet6 fe80::5055:ff:fe00:1/64 scope link noprefixroute 2026-03-10T05:28:45.247 INFO:teuthology.orchestra.run.vm01.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-10T05:28:45.247 INFO:teuthology.orchestra.run.vm01.stdout:Mon IP `192.168.123.101` is in CIDR network `192.168.123.0/24` 2026-03-10T05:28:45.247 INFO:teuthology.orchestra.run.vm01.stdout:Mon IP `192.168.123.101` is in CIDR network `192.168.123.0/24` 2026-03-10T05:28:45.247 INFO:teuthology.orchestra.run.vm01.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24'] 2026-03-10T05:28:45.248 INFO:teuthology.orchestra.run.vm01.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-10T05:28:45.248 INFO:teuthology.orchestra.run.vm01.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T05:28:46.586 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stdout 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-10T05:28:46.586 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Trying to pull quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T05:28:46.586 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Getting image source signatures 2026-03-10T05:28:46.586 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Copying blob sha256:1752b8d01aa0dd33bbe0ab24e8316174c94fbdcd5d26252e2680bba0624747a7 2026-03-10T05:28:46.586 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Copying blob sha256:8e380faede39ebd4286247457b408d979ab568aafd8389c42ec304b8cfba4e92 2026-03-10T05:28:46.586 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Copying config sha256:654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-10T05:28:46.586 INFO:teuthology.orchestra.run.vm01.stdout:/bin/podman: stderr Writing manifest to image destination 2026-03-10T05:28:46.730 INFO:teuthology.orchestra.run.vm01.stdout:ceph: stdout ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-10T05:28:46.731 INFO:teuthology.orchestra.run.vm01.stdout:Ceph version: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-10T05:28:46.731 INFO:teuthology.orchestra.run.vm01.stdout:Extracting ceph user uid/gid from container image... 2026-03-10T05:28:46.846 INFO:teuthology.orchestra.run.vm01.stdout:stat: stdout 167 167 2026-03-10T05:28:46.846 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial keys... 2026-03-10T05:28:46.961 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQAOrK9pF3OqNhAAGc75l0jfeKaerbiSfEPxuQ== 2026-03-10T05:28:47.063 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQAPrK9pvbsbAhAA8tAIZf7ihoIIOVGfZk1Tkw== 2026-03-10T05:28:47.181 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph-authtool: stdout AQAPrK9paDXgBxAAymORT/LTJdR39Nlh45UIWA== 2026-03-10T05:28:47.182 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial monmap... 2026-03-10T05:28:47.307 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T05:28:47.307 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-10T05:28:47.307 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:28:47.307 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T05:28:47.307 INFO:teuthology.orchestra.run.vm01.stdout:monmaptool for a [v2:192.168.123.101:3300,v1:192.168.123.101:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T05:28:47.307 INFO:teuthology.orchestra.run.vm01.stdout:setting min_mon_release = quincy 2026-03-10T05:28:47.307 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: set fsid to dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:28:47.307 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T05:28:47.307 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:28:47.308 INFO:teuthology.orchestra.run.vm01.stdout:Creating mon... 2026-03-10T05:28:47.513 INFO:teuthology.orchestra.run.vm01.stdout:create mon.a on 2026-03-10T05:28:47.792 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-10T05:28:47.938 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba.target → /etc/systemd/system/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba.target. 2026-03-10T05:28:47.938 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba.target → /etc/systemd/system/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba.target. 2026-03-10T05:28:48.107 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.a 2026-03-10T05:28:48.107 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to reset failed state of unit ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.a.service: Unit ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.a.service not loaded. 2026-03-10T05:28:48.257 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba.target.wants/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.a.service → /etc/systemd/system/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@.service. 2026-03-10T05:28:48.436 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-10T05:28:48.436 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to enable service . firewalld.service is not available 2026-03-10T05:28:48.436 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mon to start... 2026-03-10T05:28:48.436 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mon... 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout cluster: 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout id: dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout services: 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.148209s) 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout data: 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout pgs: 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:mon is available 2026-03-10T05:28:48.673 INFO:teuthology.orchestra.run.vm01.stdout:Assimilating anything we can from ceph.conf... 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [global] 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout fsid = dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [osd] 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-10T05:28:48.900 INFO:teuthology.orchestra.run.vm01.stdout:Generating new minimal ceph.conf... 2026-03-10T05:28:49.089 INFO:teuthology.orchestra.run.vm01.stdout:Restarting the monitor... 2026-03-10T05:28:49.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-a[47651]: 2026-03-10T05:28:49.170+0000 7f8b0539a640 -1 mon.a@0(leader) e1 *** Got Signal Terminated *** 2026-03-10T05:28:49.634 INFO:teuthology.orchestra.run.vm01.stdout:Setting public_network to 192.168.123.0/24 in mon config section 2026-03-10T05:28:49.662 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 podman[47856]: 2026-03-10 05:28:49.406307573 +0000 UTC m=+0.246715414 container died 4a59fb36a115fd23edfa81b583dca32e2461b0ae10dc941af286bdceb5af611e (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-a, OSD_FLAVOR=default, CEPH_REF=squid, ceph=True, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-10T05:28:49.662 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 podman[47856]: 2026-03-10 05:28:49.417183359 +0000 UTC m=+0.257591200 container remove 4a59fb36a115fd23edfa81b583dca32e2461b0ae10dc941af286bdceb5af611e (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-a, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, ceph=True, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-10T05:28:49.662 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 bash[47856]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-a 2026-03-10T05:28:49.662 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.a.service: Deactivated successfully. 2026-03-10T05:28:49.662 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 systemd[1]: Stopped Ceph mon.a for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:28:49.662 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 systemd[1]: Starting Ceph mon.a for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 podman[47927]: 2026-03-10 05:28:49.575573332 +0000 UTC m=+0.017352100 container create 11f18971109c816585547659c764de6b76b565ab362f65c419b1640fa99ce169 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-a, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, CEPH_REF=squid, org.label-schema.schema-version=1.0, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 podman[47927]: 2026-03-10 05:28:49.617036059 +0000 UTC m=+0.058814837 container init 11f18971109c816585547659c764de6b76b565ab362f65c419b1640fa99ce169 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-a, org.opencontainers.image.authors=Ceph Release Team , ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223) 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 podman[47927]: 2026-03-10 05:28:49.620979881 +0000 UTC m=+0.062758659 container start 11f18971109c816585547659c764de6b76b565ab362f65c419b1640fa99ce169 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-a, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, ceph=True, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 bash[47927]: 11f18971109c816585547659c764de6b76b565ab362f65c419b1640fa99ce169 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 podman[47927]: 2026-03-10 05:28:49.568003083 +0000 UTC m=+0.009781870 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 systemd[1]: Started Ceph mon.a for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: set uid:gid to 167:167 (ceph:ceph) 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 2 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: pidfile_write: ignore empty --pid-file 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: load: jerasure load: lrc 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: RocksDB version: 7.9.2 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Git sha 0 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: DB SUMMARY 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: DB Session ID: HSB0BQXIX5UIQ0LSZ0KO 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: CURRENT file: CURRENT 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: IDENTITY file: IDENTITY 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 75519 ; 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.error_if_exists: 0 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.create_if_missing: 0 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.paranoid_checks: 1 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.env: 0x56063f506dc0 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.fs: PosixFileSystem 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.info_log: 0x5606417b2700 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_file_opening_threads: 16 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.statistics: (nil) 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.use_fsync: 0 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_log_file_size: 0 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.keep_log_file_num: 1000 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.recycle_log_file_num: 0 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.allow_fallocate: 1 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.allow_mmap_reads: 0 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.allow_mmap_writes: 0 2026-03-10T05:28:49.663 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.use_direct_reads: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.create_missing_column_families: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.db_log_dir: 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.wal_dir: 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.advise_random_on_open: 1 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.db_write_buffer_size: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.write_buffer_manager: 0x5606417b7900 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.rate_limiter: (nil) 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.wal_recovery_mode: 2 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.enable_thread_tracking: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.enable_pipelined_write: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.unordered_write: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.row_cache: None 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.wal_filter: None 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.allow_ingest_behind: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.two_write_queues: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.manual_wal_flush: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.wal_compression: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.atomic_flush: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.log_readahead_size: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.best_efforts_recovery: 0 2026-03-10T05:28:49.664 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.allow_data_in_errors: 0 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.db_host_id: __hostname__ 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_background_jobs: 2 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_background_compactions: -1 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_subcompactions: 1 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_total_wal_size: 0 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T05:28:49.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_open_files: -1 2026-03-10T05:28:49.872 INFO:teuthology.orchestra.run.vm01.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-10T05:28:49.873 INFO:teuthology.orchestra.run.vm01.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:28:49.873 INFO:teuthology.orchestra.run.vm01.stdout:Creating mgr... 2026-03-10T05:28:49.873 INFO:teuthology.orchestra.run.vm01.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-10T05:28:49.874 INFO:teuthology.orchestra.run.vm01.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.bytes_per_sync: 0 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compaction_readahead_size: 0 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_background_flushes: -1 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Compression algorithms supported: 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: kZSTD supported: 0 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: kXpressCompression supported: 0 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: kBZip2Compression supported: 0 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: kLZ4Compression supported: 1 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: kZlibCompression supported: 1 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: kLZ4HCCompression supported: 1 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: kSnappyCompression supported: 1 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.merge_operator: 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compaction_filter: None 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compaction_filter_factory: None 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.sst_partitioner_factory: None 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5606417b2640) 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout: cache_index_and_filter_blocks: 1 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout: pin_top_level_index_and_filter: 1 2026-03-10T05:28:49.946 INFO:journalctl@ceph.mon.a.vm01.stdout: index_type: 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: data_block_index_type: 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: index_shortening: 1 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: checksum: 4 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: no_block_cache: 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: block_cache: 0x5606417d7350 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: block_cache_name: BinnedLRUCache 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: block_cache_options: 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: capacity : 536870912 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: num_shard_bits : 4 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: strict_capacity_limit : 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: high_pri_pool_ratio: 0.000 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: block_cache_compressed: (nil) 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: persistent_cache: (nil) 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: block_size: 4096 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: block_size_deviation: 10 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: block_restart_interval: 16 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: index_block_restart_interval: 1 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: metadata_block_size: 4096 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: partition_filters: 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: use_delta_encoding: 1 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: filter_policy: bloomfilter 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: whole_key_filtering: 1 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: verify_compression: 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: read_amp_bytes_per_bit: 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: format_version: 5 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: enable_index_compression: 1 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: block_align: 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: max_auto_readahead_size: 262144 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: prepopulate_block_cache: 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: initial_auto_readahead_size: 8192 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout: num_file_reads_for_auto_readahead: 2 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.write_buffer_size: 33554432 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_write_buffer_number: 2 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compression: NoCompression 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.bottommost_compression: Disabled 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.prefix_extractor: nullptr 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.num_levels: 7 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T05:28:49.947 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compression_opts.level: 32767 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compression_opts.strategy: 0 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compression_opts.enabled: false 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.target_file_size_base: 67108864 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.arena_block_size: 1048576 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.disable_auto_compactions: 0 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.inplace_update_support: 0 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.bloom_locality: 0 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.max_successive_merges: 0 2026-03-10T05:28:49.948 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.paranoid_file_checks: 0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.force_consistency_checks: 1 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.report_bg_io_stats: 0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.ttl: 2592000 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.enable_blob_files: false 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.min_blob_size: 0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.blob_file_size: 268435456 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.blob_file_starting_level: 0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 2f1951d1-bb8f-41a7-94b7-dce84309b069 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773120529665770, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773120529669356, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 72589, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 225, "table_properties": {"data_size": 70864, "index_size": 178, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9693, "raw_average_key_size": 49, "raw_value_size": 65370, "raw_average_value_size": 333, "num_data_blocks": 8, "num_entries": 196, "num_filter_entries": 196, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773120529, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "2f1951d1-bb8f-41a7-94b7-dce84309b069", "db_session_id": "HSB0BQXIX5UIQ0LSZ0KO", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773120529669444, "job": 1, "event": "recovery_finished"} 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x5606417d8e00 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: DB pointer 0x5606418f2000 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: ** DB Stats ** 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: ** Compaction Stats [default] ** 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: L0 2/0 72.74 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 20.8 0.00 0.00 1 0.003 0 0 0.0 0.0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Sum 2/0 72.74 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 20.8 0.00 0.00 1 0.003 0 0 0.0 0.0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 20.8 0.00 0.00 1 0.003 0 0 0.0 0.0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: ** Compaction Stats [default] ** 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 20.8 0.00 0.00 1 0.003 0 0 0.0 0.0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Cumulative compaction: 0.00 GB write, 4.12 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T05:28:49.949 INFO:journalctl@ceph.mon.a.vm01.stdout: Interval compaction: 0.00 GB write, 4.12 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout: Block cache BinnedLRUCache@0x5606417d7350#2 capacity: 512.00 MB usage: 26.19 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 7e-06 secs_since: 0 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout: Block cache entry stats(count,size,portion): DataBlock(3,25.11 KB,0.00478923%) FilterBlock(2,0.70 KB,0.00013411%) IndexBlock(2,0.38 KB,7.15256e-05%) Misc(1,0.00 KB,0%) 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: starting mon.a rank 0 at public addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] at bind addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: mon.a@-1(???) e1 preinit fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: mon.a@-1(???).mds e1 new map 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: mon.a@-1(???).mds e1 print_map 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout: e1 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout: btime 2026-03-10T05:28:48:468632+0000 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout: legacy client fscid: -1 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout: 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout: No filesystems configured 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: mon.a@-1(???).osd e1 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: monmap epoch 1 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: last_changed 2026-03-10T05:28:47.264782+0000 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: created 2026-03-10T05:28:47.264782+0000 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: min_mon_release 19 (squid) 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: election_strategy: 1 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: fsmap 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: osdmap e1: 0 total, 0 up, 0 in 2026-03-10T05:28:49.950 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:49 vm01 ceph-mon[47941]: mgrmap e1: no daemons active 2026-03-10T05:28:50.032 INFO:teuthology.orchestra.run.vm01.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mgr.x 2026-03-10T05:28:50.032 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Failed to reset failed state of unit ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mgr.x.service: Unit ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mgr.x.service not loaded. 2026-03-10T05:28:50.169 INFO:teuthology.orchestra.run.vm01.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba.target.wants/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mgr.x.service → /etc/systemd/system/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@.service. 2026-03-10T05:28:50.306 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:50 vm01 systemd[1]: Starting Ceph mgr.x for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:28:50.355 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-10T05:28:50.355 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to enable service . firewalld.service is not available 2026-03-10T05:28:50.355 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-10T05:28:50.355 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-10T05:28:50.355 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr to start... 2026-03-10T05:28:50.355 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr... 2026-03-10T05:28:50.589 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:50 vm01 podman[48139]: 2026-03-10 05:28:50.304052173 +0000 UTC m=+0.021337859 container create 031e5c93146d7d6adf57c8a41daab78090cc8ed619b7bb2d9c824ff023deed7f (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_REF=squid, org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0) 2026-03-10T05:28:50.589 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:50 vm01 podman[48139]: 2026-03-10 05:28:50.34112818 +0000 UTC m=+0.058413866 container init 031e5c93146d7d6adf57c8a41daab78090cc8ed619b7bb2d9c824ff023deed7f (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x, OSD_FLAVOR=default, io.buildah.version=1.41.3, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, CEPH_REF=squid, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-10T05:28:50.589 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:50 vm01 podman[48139]: 2026-03-10 05:28:50.344001508 +0000 UTC m=+0.061287194 container start 031e5c93146d7d6adf57c8a41daab78090cc8ed619b7bb2d9c824ff023deed7f (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, CEPH_REF=squid, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True) 2026-03-10T05:28:50.589 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:50 vm01 bash[48139]: 031e5c93146d7d6adf57c8a41daab78090cc8ed619b7bb2d9c824ff023deed7f 2026-03-10T05:28:50.589 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:50 vm01 podman[48139]: 2026-03-10 05:28:50.295604633 +0000 UTC m=+0.012890328 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:28:50.589 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:50 vm01 systemd[1]: Started Ceph mgr.x for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:28:50.589 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:50.467+0000 7f298d0df140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T05:28:50.589 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:50.520+0000 7f298d0df140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T05:28:50.668 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:28:50.669 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T05:28:50.669 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba", 2026-03-10T05:28:50.669 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T05:28:50.669 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 0, 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:50.670 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T05:28:48:468632+0000", 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T05:28:48.470472+0000", 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T05:28:50.671 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (1/15)... 2026-03-10T05:28:50.991 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:50 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1945772495' entity='client.admin' 2026-03-10T05:28:50.991 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:50 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2434590773' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T05:28:51.333 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:50.988+0000 7f298d0df140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T05:28:51.333 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:51.330+0000 7f298d0df140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T05:28:51.588 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T05:28:51.588 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T05:28:51.588 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: from numpy import show_config as show_numpy_config 2026-03-10T05:28:51.588 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:51.429+0000 7f298d0df140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T05:28:51.588 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:51.468+0000 7f298d0df140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T05:28:51.588 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:51.544+0000 7f298d0df140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T05:28:52.377 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:52.087+0000 7f298d0df140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T05:28:52.377 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:52.202+0000 7f298d0df140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T05:28:52.377 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:52.245+0000 7f298d0df140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T05:28:52.377 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:52.282+0000 7f298d0df140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T05:28:52.377 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:52.327+0000 7f298d0df140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T05:28:52.377 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:52.374+0000 7f298d0df140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T05:28:52.637 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:52.564+0000 7f298d0df140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T05:28:52.638 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:52.633+0000 7f298d0df140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T05:28:52.915 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:28:52.915 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T05:28:52.915 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba", 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T05:28:52.916 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T05:28:48:468632+0000", 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T05:28:48.470472+0000", 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T05:28:52.917 INFO:teuthology.orchestra.run.vm01.stdout:mgr not available, waiting (2/15)... 2026-03-10T05:28:53.211 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:52.895+0000 7f298d0df140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T05:28:53.211 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4163284764' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T05:28:53.504 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:53.208+0000 7f298d0df140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T05:28:53.504 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:53.248+0000 7f298d0df140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T05:28:53.504 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:53.292+0000 7f298d0df140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T05:28:53.504 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:53.374+0000 7f298d0df140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T05:28:53.504 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:53.414+0000 7f298d0df140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T05:28:53.778 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:53.501+0000 7f298d0df140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T05:28:53.778 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:53.624+0000 7f298d0df140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:53.775+0000 7f298d0df140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:53.816+0000 7f298d0df140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:53 vm01 ceph-mon[47941]: Activating manager daemon x 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:53 vm01 ceph-mon[47941]: mgrmap e2: x(active, starting, since 0.00509777s) 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:53 vm01 ceph-mon[47941]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:53 vm01 ceph-mon[47941]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:53 vm01 ceph-mon[47941]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:53 vm01 ceph-mon[47941]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:53 vm01 ceph-mon[47941]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:53 vm01 ceph-mon[47941]: Manager daemon x is now available 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:53 vm01 ceph-mon[47941]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"}]: dispatch 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:53 vm01 ceph-mon[47941]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:53 vm01 ceph-mon[47941]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"}]: dispatch 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:53 vm01 ceph-mon[47941]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' 2026-03-10T05:28:54.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:53 vm01 ceph-mon[47941]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' 2026-03-10T05:28:55.237 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:28:55.237 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T05:28:55.237 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsid": "dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba", 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 0 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T05:28:48:468632+0000", 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T05:28:55.238 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ], 2026-03-10T05:28:55.239 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T05:28:55.239 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:55.239 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T05:28:55.239 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T05:28:55.239 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T05:28:48.470472+0000", 2026-03-10T05:28:55.239 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T05:28:55.239 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout }, 2026-03-10T05:28:55.239 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T05:28:55.239 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T05:28:55.239 INFO:teuthology.orchestra.run.vm01.stdout:mgr is available 2026-03-10T05:28:55.537 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [global] 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout fsid = dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.101:3300,v1:192.168.123.101:6789] 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout [osd] 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-10T05:28:55.538 INFO:teuthology.orchestra.run.vm01.stdout:Enabling cephadm module... 2026-03-10T05:28:55.829 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:55 vm01 ceph-mon[47941]: mgrmap e3: x(active, since 1.01059s) 2026-03-10T05:28:55.829 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:55 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3770986770' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T05:28:55.829 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:55 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/759300726' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T05:28:55.829 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:55 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/888769153' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T05:28:56.089 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: ignoring --setuser ceph since I am not root 2026-03-10T05:28:56.089 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: ignoring --setgroup ceph since I am not root 2026-03-10T05:28:56.089 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:55.986+0000 7feff1c34140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T05:28:56.089 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:56.040+0000 7feff1c34140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T05:28:56.197 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T05:28:56.198 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 4, 2026-03-10T05:28:56.198 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T05:28:56.198 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "active_name": "x", 2026-03-10T05:28:56.198 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-10T05:28:56.198 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T05:28:56.198 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for the mgr to restart... 2026-03-10T05:28:56.198 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr epoch 4... 2026-03-10T05:28:56.834 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:56.525+0000 7feff1c34140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T05:28:57.088 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:56.892+0000 7feff1c34140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T05:28:57.089 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T05:28:57.089 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T05:28:57.089 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: from numpy import show_config as show_numpy_config 2026-03-10T05:28:57.089 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:56.988+0000 7feff1c34140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T05:28:57.089 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:57.033+0000 7feff1c34140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T05:28:57.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:56 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/888769153' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T05:28:57.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:56 vm01 ceph-mon[47941]: mgrmap e4: x(active, since 2s) 2026-03-10T05:28:57.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:56 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/901160325' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T05:28:57.588 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:57.117+0000 7feff1c34140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T05:28:57.927 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:57.666+0000 7feff1c34140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T05:28:57.927 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:57.790+0000 7feff1c34140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T05:28:57.927 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:57.839+0000 7feff1c34140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T05:28:57.927 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:57.880+0000 7feff1c34140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T05:28:58.210 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:57.924+0000 7feff1c34140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T05:28:58.210 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:57.963+0000 7feff1c34140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T05:28:58.210 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:58.147+0000 7feff1c34140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T05:28:58.210 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:58.207+0000 7feff1c34140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T05:28:58.588 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:58.446+0000 7feff1c34140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T05:28:59.050 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:58.755+0000 7feff1c34140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T05:28:59.050 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:58.794+0000 7feff1c34140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T05:28:59.050 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:58.838+0000 7feff1c34140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T05:28:59.050 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:58.920+0000 7feff1c34140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T05:28:59.050 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:58.961+0000 7feff1c34140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T05:28:59.315 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:59.046+0000 7feff1c34140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T05:28:59.315 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:59.166+0000 7feff1c34140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T05:28:59.588 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:59.312+0000 7feff1c34140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T05:28:59.588 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:28:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:28:59.350+0000 7feff1c34140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T05:29:00.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:59 vm01 ceph-mon[47941]: Active manager daemon x restarted 2026-03-10T05:29:00.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:59 vm01 ceph-mon[47941]: Activating manager daemon x 2026-03-10T05:29:00.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:59 vm01 ceph-mon[47941]: osdmap e2: 0 total, 0 up, 0 in 2026-03-10T05:29:00.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:59 vm01 ceph-mon[47941]: mgrmap e5: x(active, starting, since 0.560737s) 2026-03-10T05:29:00.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:59 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:29:00.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:59 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-10T05:29:00.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:59 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T05:29:00.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:59 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T05:29:00.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:59 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T05:29:00.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:59 vm01 ceph-mon[47941]: Manager daemon x is now available 2026-03-10T05:29:00.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:59 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:00.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:59 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:00.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:59 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:00.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:28:59 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:00.967 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T05:29:00.967 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 5, 2026-03-10T05:29:00.967 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-10T05:29:00.967 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T05:29:00.967 INFO:teuthology.orchestra.run.vm01.stdout:mgr epoch 4 is available 2026-03-10T05:29:00.967 INFO:teuthology.orchestra.run.vm01.stdout:Setting orchestrator backend to cephadm... 2026-03-10T05:29:01.242 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:00 vm01 ceph-mon[47941]: Found migration_current of "None". Setting to last migration. 2026-03-10T05:29:01.242 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:00 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"}]: dispatch 2026-03-10T05:29:01.242 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:00 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"}]: dispatch 2026-03-10T05:29:01.242 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:00 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:01.242 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:00 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:01.242 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:00 vm01 ceph-mon[47941]: mgrmap e6: x(active, since 1.56514s) 2026-03-10T05:29:01.562 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-10T05:29:01.563 INFO:teuthology.orchestra.run.vm01.stdout:Generating ssh key... 2026-03-10T05:29:02.063 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: Generating public/private ed25519 key pair. 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: Your identification has been saved in /tmp/tmpreqaban0/key 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: Your public key has been saved in /tmp/tmpreqaban0/key.pub 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: The key fingerprint is: 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: SHA256:EnjVx1nX+JOkHPP6DWZPqxcni+h6QXN+T6HPn+B5kYg ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: The key's randomart image is: 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: +--[ED25519 256]--+ 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: | .. . o..o| 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: | . . . +o.o.| 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: | . o .. *..| 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: | . . o .o =.| 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: | . S. +. + +| 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: | . .E.+*++| 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: | o *+O*| 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: | o o +B=| 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: | .+. ++.o| 2026-03-10T05:29:02.064 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: +----[SHA256]-----+ 2026-03-10T05:29:02.114 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEDxYAotktVJ3k8PJ6bt9y7XYFgt+9Cqsa/txkncEH5g ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:02.114 INFO:teuthology.orchestra.run.vm01.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-10T05:29:02.114 INFO:teuthology.orchestra.run.vm01.stdout:Adding key to root@localhost authorized_keys... 2026-03-10T05:29:02.115 INFO:teuthology.orchestra.run.vm01.stdout:Adding host vm01... 2026-03-10T05:29:02.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T05:29:02.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T05:29:02.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: [10/Mar/2026:05:29:01] ENGINE Bus STARTING 2026-03-10T05:29:02.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: [10/Mar/2026:05:29:01] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T05:29:02.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:02.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:02.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:02.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: [10/Mar/2026:05:29:01] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T05:29:02.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: [10/Mar/2026:05:29:01] ENGINE Bus STARTED 2026-03-10T05:29:02.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: [10/Mar/2026:05:29:01] ENGINE Client ('192.168.123.101', 42554) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T05:29:02.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:02.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:02.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:02.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: Generating ssh key... 2026-03-10T05:29:02.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:02.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:02 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:03 vm01 ceph-mon[47941]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:03 vm01 ceph-mon[47941]: mgrmap e7: x(active, since 2s) 2026-03-10T05:29:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:03 vm01 ceph-mon[47941]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm01", "addr": "192.168.123.101", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:03 vm01 ceph-mon[47941]: Deploying cephadm binary to vm01 2026-03-10T05:29:04.297 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Added host 'vm01' with addr '192.168.123.101' 2026-03-10T05:29:04.297 INFO:teuthology.orchestra.run.vm01.stdout:Deploying unmanaged mon service... 2026-03-10T05:29:04.585 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-10T05:29:04.585 INFO:teuthology.orchestra.run.vm01.stdout:Deploying unmanaged mgr service... 2026-03-10T05:29:04.896 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-10T05:29:05.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:05 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:05.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:05 vm01 ceph-mon[47941]: Added host vm01 2026-03-10T05:29:05.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:05 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:05.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:05 vm01 ceph-mon[47941]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:05.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:05 vm01 ceph-mon[47941]: Saving service mon spec with placement count:5 2026-03-10T05:29:05.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:05 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:05.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:05 vm01 ceph-mon[47941]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:05.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:05 vm01 ceph-mon[47941]: Saving service mgr spec with placement count:2 2026-03-10T05:29:05.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:05 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:05.424 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:05 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2407811369' entity='client.admin' 2026-03-10T05:29:05.463 INFO:teuthology.orchestra.run.vm01.stdout:Enabling the dashboard module... 2026-03-10T05:29:06.649 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:06 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4184714749' entity='client.admin' 2026-03-10T05:29:06.649 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:06 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2335065685' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T05:29:06.649 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:06 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:06.649 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:06 vm01 ceph-mon[47941]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:06.649 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:06 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: ignoring --setuser ceph since I am not root 2026-03-10T05:29:06.649 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:06 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: ignoring --setgroup ceph since I am not root 2026-03-10T05:29:06.649 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:06 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:06.609+0000 7fa5fb16f140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T05:29:06.888 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T05:29:06.888 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "epoch": 8, 2026-03-10T05:29:06.888 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T05:29:06.888 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "active_name": "x", 2026-03-10T05:29:06.888 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-10T05:29:06.888 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T05:29:06.888 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for the mgr to restart... 2026-03-10T05:29:06.888 INFO:teuthology.orchestra.run.vm01.stdout:Waiting for mgr epoch 8... 2026-03-10T05:29:06.911 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:06 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:06.675+0000 7fa5fb16f140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T05:29:07.338 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:07 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:07.141+0000 7fa5fb16f140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T05:29:07.690 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:07 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2335065685' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T05:29:07.690 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:07 vm01 ceph-mon[47941]: mgrmap e8: x(active, since 7s) 2026-03-10T05:29:07.691 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:07 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2196795635' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T05:29:07.691 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:07 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:07.481+0000 7fa5fb16f140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T05:29:07.691 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:07 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T05:29:07.691 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:07 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T05:29:07.691 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:07 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: from numpy import show_config as show_numpy_config 2026-03-10T05:29:07.691 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:07 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:07.573+0000 7fa5fb16f140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T05:29:07.691 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:07 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:07.611+0000 7fa5fb16f140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T05:29:07.691 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:07 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:07.688+0000 7fa5fb16f140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T05:29:08.510 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:08.223+0000 7fa5fb16f140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T05:29:08.510 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:08.346+0000 7fa5fb16f140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T05:29:08.510 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:08.388+0000 7fa5fb16f140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T05:29:08.510 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:08.425+0000 7fa5fb16f140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T05:29:08.510 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:08.468+0000 7fa5fb16f140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T05:29:08.838 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:08.507+0000 7fa5fb16f140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T05:29:08.838 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:08.692+0000 7fa5fb16f140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T05:29:08.838 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:08.747+0000 7fa5fb16f140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T05:29:09.258 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:08.970+0000 7fa5fb16f140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T05:29:09.536 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:09 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:09.255+0000 7fa5fb16f140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T05:29:09.537 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:09 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:09.293+0000 7fa5fb16f140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T05:29:09.537 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:09 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:09.336+0000 7fa5fb16f140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T05:29:09.537 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:09 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:09.415+0000 7fa5fb16f140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T05:29:09.537 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:09 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:09.453+0000 7fa5fb16f140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T05:29:09.790 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:09 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:09.533+0000 7fa5fb16f140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T05:29:09.790 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:09 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:09.649+0000 7fa5fb16f140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T05:29:10.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:09 vm01 ceph-mon[47941]: Active manager daemon x restarted 2026-03-10T05:29:10.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:09 vm01 ceph-mon[47941]: Activating manager daemon x 2026-03-10T05:29:10.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:09 vm01 ceph-mon[47941]: osdmap e3: 0 total, 0 up, 0 in 2026-03-10T05:29:10.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:09 vm01 ceph-mon[47941]: mgrmap e9: x(active, starting, since 0.00692574s) 2026-03-10T05:29:10.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:09 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:29:10.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:09 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-10T05:29:10.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:09 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T05:29:10.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:09 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T05:29:10.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:09 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T05:29:10.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:09 vm01 ceph-mon[47941]: Manager daemon x is now available 2026-03-10T05:29:10.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:09 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:10.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:09 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"}]: dispatch 2026-03-10T05:29:10.089 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:09 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:09.787+0000 7fa5fb16f140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T05:29:10.089 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:09 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:09.825+0000 7fa5fb16f140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T05:29:10.900 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"}]: dispatch 2026-03-10T05:29:10.900 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:10.901 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:10 vm01 ceph-mon[47941]: mgrmap e10: x(active, since 1.00968s) 2026-03-10T05:29:10.901 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout { 2026-03-10T05:29:10.901 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 10, 2026-03-10T05:29:10.901 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-10T05:29:10.901 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout } 2026-03-10T05:29:10.901 INFO:teuthology.orchestra.run.vm01.stdout:mgr epoch 8 is available 2026-03-10T05:29:10.901 INFO:teuthology.orchestra.run.vm01.stdout:Generating a dashboard self-signed certificate... 2026-03-10T05:29:11.279 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-10T05:29:11.279 INFO:teuthology.orchestra.run.vm01.stdout:Creating initial admin user... 2026-03-10T05:29:11.701 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$NKC6hd4FVZWs8tnjzZlGk.lNyVodya1Fec3rKCCDeTsoRhYTsEmqO", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773120551, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-10T05:29:11.702 INFO:teuthology.orchestra.run.vm01.stdout:Fetching dashboard port number... 2026-03-10T05:29:11.974 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stdout 8443 2026-03-10T05:29:11.974 INFO:teuthology.orchestra.run.vm01.stdout:firewalld does not appear to be present 2026-03-10T05:29:11.974 INFO:teuthology.orchestra.run.vm01.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-10T05:29:11.976 INFO:teuthology.orchestra.run.vm01.stdout:Ceph Dashboard is now available at: 2026-03-10T05:29:11.976 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:29:11.976 INFO:teuthology.orchestra.run.vm01.stdout: URL: https://vm01.local:8443/ 2026-03-10T05:29:11.976 INFO:teuthology.orchestra.run.vm01.stdout: User: admin 2026-03-10T05:29:11.976 INFO:teuthology.orchestra.run.vm01.stdout: Password: 9goh7weneb 2026-03-10T05:29:11.976 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:29:11.976 INFO:teuthology.orchestra.run.vm01.stdout:Saving cluster configuration to /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config directory 2026-03-10T05:29:12.263 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:11 vm01 ceph-mon[47941]: [10/Mar/2026:05:29:10] ENGINE Bus STARTING 2026-03-10T05:29:12.263 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:11 vm01 ceph-mon[47941]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T05:29:12.263 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:11 vm01 ceph-mon[47941]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T05:29:12.263 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:11 vm01 ceph-mon[47941]: [10/Mar/2026:05:29:10] ENGINE Client ('192.168.123.101', 45130) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T05:29:12.263 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:11 vm01 ceph-mon[47941]: [10/Mar/2026:05:29:10] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T05:29:12.263 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:11 vm01 ceph-mon[47941]: [10/Mar/2026:05:29:10] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T05:29:12.263 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:11 vm01 ceph-mon[47941]: [10/Mar/2026:05:29:10] ENGINE Bus STARTED 2026-03-10T05:29:12.263 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:12.263 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:11 vm01 ceph-mon[47941]: from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:12.263 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:12.263 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:12.263 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:11 vm01 ceph-mon[47941]: from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:12.264 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:12.264 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:11 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3707339119' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout:Or, if you are only running a single cluster on this host: 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout: ceph telemetry on 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout:For more information see: 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:29:12.312 INFO:teuthology.orchestra.run.vm01.stdout:Bootstrap complete. 2026-03-10T05:29:12.347 INFO:tasks.cephadm:Fetching config... 2026-03-10T05:29:12.347 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:29:12.347 DEBUG:teuthology.orchestra.run.vm01:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-10T05:29:12.375 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-10T05:29:12.376 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:29:12.376 DEBUG:teuthology.orchestra.run.vm01:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-10T05:29:12.449 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-10T05:29:12.449 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:29:12.449 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/keyring of=/dev/stdout 2026-03-10T05:29:12.519 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-10T05:29:12.519 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:29:12.519 DEBUG:teuthology.orchestra.run.vm01:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-10T05:29:12.580 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-10T05:29:12.580 DEBUG:teuthology.orchestra.run.vm01:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEDxYAotktVJ3k8PJ6bt9y7XYFgt+9Cqsa/txkncEH5g ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T05:29:12.681 INFO:teuthology.orchestra.run.vm01.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEDxYAotktVJ3k8PJ6bt9y7XYFgt+9Cqsa/txkncEH5g ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:12.691 DEBUG:teuthology.orchestra.run.vm02:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEDxYAotktVJ3k8PJ6bt9y7XYFgt+9Cqsa/txkncEH5g ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T05:29:12.726 INFO:teuthology.orchestra.run.vm02.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEDxYAotktVJ3k8PJ6bt9y7XYFgt+9Cqsa/txkncEH5g ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:12.740 DEBUG:teuthology.orchestra.run.vm05:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEDxYAotktVJ3k8PJ6bt9y7XYFgt+9Cqsa/txkncEH5g ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T05:29:12.775 INFO:teuthology.orchestra.run.vm05.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIEDxYAotktVJ3k8PJ6bt9y7XYFgt+9Cqsa/txkncEH5g ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:12.784 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-10T05:29:12.957 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:29:13.571 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-10T05:29:13.571 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-10T05:29:13.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:13 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3253245897' entity='client.admin' 2026-03-10T05:29:13.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:13 vm01 ceph-mon[47941]: mgrmap e11: x(active, since 2s) 2026-03-10T05:29:13.765 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:29:14.194 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm02 2026-03-10T05:29:14.194 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:29:14.194 DEBUG:teuthology.orchestra.run.vm02:> dd of=/etc/ceph/ceph.conf 2026-03-10T05:29:14.214 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:29:14.214 DEBUG:teuthology.orchestra.run.vm02:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:29:14.274 INFO:tasks.cephadm:Adding host vm02 to orchestrator... 2026-03-10T05:29:14.275 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch host add vm02 2026-03-10T05:29:14.475 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:29:14.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:14 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1595755798' entity='client.admin' 2026-03-10T05:29:14.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:14.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:14.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:29:14.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:14.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:14.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:14.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:14.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:14.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:14.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:15.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:15 vm01 ceph-mon[47941]: from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:15.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:15 vm01 ceph-mon[47941]: Updating vm01:/etc/ceph/ceph.conf 2026-03-10T05:29:15.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:15 vm01 ceph-mon[47941]: Updating vm01:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:15.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:15 vm01 ceph-mon[47941]: Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:29:15.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:15.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:15.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:16.357 INFO:teuthology.orchestra.run.vm01.stdout:Added host 'vm02' with addr '192.168.123.102' 2026-03-10T05:29:16.406 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch host ls --format=json 2026-03-10T05:29:16.604 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:29:16.629 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:16 vm01 ceph-mon[47941]: Updating vm01:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.client.admin.keyring 2026-03-10T05:29:16.629 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:16 vm01 ceph-mon[47941]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:16.629 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:16 vm01 ceph-mon[47941]: Deploying cephadm binary to vm02 2026-03-10T05:29:16.629 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:16 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:16.629 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:16 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:16.838 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:29:16.838 INFO:teuthology.orchestra.run.vm01.stdout:[{"addr": "192.168.123.101", "hostname": "vm01", "labels": [], "status": ""}, {"addr": "192.168.123.102", "hostname": "vm02", "labels": [], "status": ""}] 2026-03-10T05:29:16.915 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm05 2026-03-10T05:29:16.915 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T05:29:16.915 DEBUG:teuthology.orchestra.run.vm05:> dd of=/etc/ceph/ceph.conf 2026-03-10T05:29:16.930 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T05:29:16.930 DEBUG:teuthology.orchestra.run.vm05:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:29:16.988 INFO:tasks.cephadm:Adding host vm05 to orchestrator... 2026-03-10T05:29:16.988 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch host add vm05 2026-03-10T05:29:17.159 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:29:17.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:17 vm01 ceph-mon[47941]: Added host vm02 2026-03-10T05:29:17.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:17 vm01 ceph-mon[47941]: mgrmap e12: x(active, since 6s) 2026-03-10T05:29:17.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:17.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:18.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:18 vm01 ceph-mon[47941]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:29:18.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:18 vm01 ceph-mon[47941]: from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm05", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:18.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:18 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:18.940 INFO:teuthology.orchestra.run.vm01.stdout:Added host 'vm05' with addr '192.168.123.105' 2026-03-10T05:29:18.998 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch host ls --format=json 2026-03-10T05:29:19.181 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:29:19.416 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:29:19.416 INFO:teuthology.orchestra.run.vm01.stdout:[{"addr": "192.168.123.101", "hostname": "vm01", "labels": [], "status": ""}, {"addr": "192.168.123.102", "hostname": "vm02", "labels": [], "status": ""}, {"addr": "192.168.123.105", "hostname": "vm05", "labels": [], "status": ""}] 2026-03-10T05:29:19.481 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-10T05:29:19.481 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd crush tunables default 2026-03-10T05:29:19.661 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:29:19.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:19 vm01 ceph-mon[47941]: Deploying cephadm binary to vm05 2026-03-10T05:29:19.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:19 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:19.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:19 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:19.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:19 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:19.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:19 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:19.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:19 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:19.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:19 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:29:19.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:19 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:19.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:19 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:19.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:19 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:19.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:19 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:19.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:19 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:20.735 INFO:teuthology.orchestra.run.vm01.stderr:adjusted tunables profile to default 2026-03-10T05:29:20.772 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:20 vm01 ceph-mon[47941]: Added host vm05 2026-03-10T05:29:20.772 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:20 vm01 ceph-mon[47941]: Updating vm02:/etc/ceph/ceph.conf 2026-03-10T05:29:20.772 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:20 vm01 ceph-mon[47941]: Updating vm02:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:20.772 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:20 vm01 ceph-mon[47941]: Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:29:20.772 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:20 vm01 ceph-mon[47941]: Updating vm02:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.client.admin.keyring 2026-03-10T05:29:20.772 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:20 vm01 ceph-mon[47941]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:29:20.773 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:20 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:20.773 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:20 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:20.773 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:20 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1345241876' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T05:29:20.773 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:20 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:20.792 INFO:tasks.cephadm:Adding mon.a on vm01 2026-03-10T05:29:20.793 INFO:tasks.cephadm:Adding mon.b on vm02 2026-03-10T05:29:20.793 INFO:tasks.cephadm:Adding mon.c on vm05 2026-03-10T05:29:20.793 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch apply mon '3;vm01:192.168.123.101=a;vm02:192.168.123.102=b;vm05:192.168.123.105=c' 2026-03-10T05:29:20.990 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:29:21.032 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:29:21.308 INFO:teuthology.orchestra.run.vm05.stdout:Scheduled mon update... 2026-03-10T05:29:21.387 DEBUG:teuthology.orchestra.run.vm02:mon.b> sudo journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.b.service 2026-03-10T05:29:21.389 DEBUG:teuthology.orchestra.run.vm05:mon.c> sudo journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.c.service 2026-03-10T05:29:21.390 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-10T05:29:21.390 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph mon dump -f json 2026-03-10T05:29:21.667 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:29:21.707 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T05:29:22.007 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:29:22.008 INFO:teuthology.orchestra.run.vm05.stdout:{"epoch":1,"fsid":"dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba","modified":"2026-03-10T05:28:47.264782Z","created":"2026-03-10T05:28:47.264782Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:29:22.008 INFO:teuthology.orchestra.run.vm05.stderr:dumped monmap epoch 1 2026-03-10T05:29:22.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:21 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1345241876' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T05:29:22.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:21 vm01 ceph-mon[47941]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T05:29:22.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:21 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:22.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:21 vm01 ceph-mon[47941]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm01:192.168.123.101=a;vm02:192.168.123.102=b;vm05:192.168.123.105=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:22.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:21 vm01 ceph-mon[47941]: Saving service mon spec with placement vm01:192.168.123.101=a;vm02:192.168.123.102=b;vm05:192.168.123.105=c;count:3 2026-03-10T05:29:22.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:21 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:23.077 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-10T05:29:23.077 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph mon dump -f json 2026-03-10T05:29:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:22 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1762770644' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:29:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:23.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:23.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:29:23.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:23.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:23.286 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:23.582 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:29:23.583 INFO:teuthology.orchestra.run.vm05.stdout:{"epoch":1,"fsid":"dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba","modified":"2026-03-10T05:28:47.264782Z","created":"2026-03-10T05:28:47.264782Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T05:29:23.583 INFO:teuthology.orchestra.run.vm05.stderr:dumped monmap epoch 1 2026-03-10T05:29:24.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:24 vm01 ceph-mon[47941]: Updating vm05:/etc/ceph/ceph.conf 2026-03-10T05:29:24.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:24 vm01 ceph-mon[47941]: Updating vm05:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:24.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:24 vm01 ceph-mon[47941]: Updating vm05:/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:29:24.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:24 vm01 ceph-mon[47941]: Updating vm05:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.client.admin.keyring 2026-03-10T05:29:24.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:24 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:24.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:24 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:24.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:24 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:24.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:24 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:29:24.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:24 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:24.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:24 vm01 ceph-mon[47941]: Deploying daemon mon.c on vm05 2026-03-10T05:29:24.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:24 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/748299835' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:29:24.655 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-10T05:29:24.655 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph mon dump -f json 2026-03-10T05:29:24.872 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.c@-1(synchronizing).mds e1 new map 2026-03-10T05:29:24.895 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.c/config 2026-03-10T05:29:25.308 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.c@-1(synchronizing).mds e1 print_map 2026-03-10T05:29:25.308 INFO:journalctl@ceph.mon.c.vm05.stdout: e1 2026-03-10T05:29:25.308 INFO:journalctl@ceph.mon.c.vm05.stdout: btime 2026-03-10T05:28:48:468632+0000 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout: legacy client fscid: -1 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout: 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout: No filesystems configured 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.c@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.c@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.c@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.c@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.c@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.c@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.c@-1(synchronizing).osd e4 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mkfs dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: monmap epoch 1 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: last_changed 2026-03-10T05:28:47.264782+0000 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: created 2026-03-10T05:28:47.264782+0000 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: min_mon_release 19 (squid) 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: election_strategy: 1 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: fsmap 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: osdmap e1: 0 total, 0 up, 0 in 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mgrmap e1: no daemons active 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1945772495' entity='client.admin' 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2434590773' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4163284764' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Activating manager daemon x 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mgrmap e2: x(active, starting, since 0.00509777s) 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Manager daemon x is now available 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"}]: dispatch 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"}]: dispatch 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14100 192.168.123.101:0/4290329623' entity='mgr.x' 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mgrmap e3: x(active, since 1.01059s) 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3770986770' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/759300726' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/888769153' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/888769153' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mgrmap e4: x(active, since 2s) 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/901160325' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Active manager daemon x restarted 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Activating manager daemon x 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: osdmap e2: 0 total, 0 up, 0 in 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mgrmap e5: x(active, starting, since 0.560737s) 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:29:25.309 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Manager daemon x is now available 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Found migration_current of "None". Setting to last migration. 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mgrmap e6: x(active, since 1.56514s) 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: [10/Mar/2026:05:29:01] ENGINE Bus STARTING 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: [10/Mar/2026:05:29:01] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: [10/Mar/2026:05:29:01] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: [10/Mar/2026:05:29:01] ENGINE Bus STARTED 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: [10/Mar/2026:05:29:01] ENGINE Client ('192.168.123.101', 42554) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Generating ssh key... 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mgrmap e7: x(active, since 2s) 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm01", "addr": "192.168.123.101", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Deploying cephadm binary to vm01 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Added host vm01 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Saving service mon spec with placement count:5 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:25.310 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Saving service mgr spec with placement count:2 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2407811369' entity='client.admin' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4184714749' entity='client.admin' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2335065685' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14118 192.168.123.101:0/4293839417' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2335065685' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mgrmap e8: x(active, since 7s) 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2196795635' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Active manager daemon x restarted 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Activating manager daemon x 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: osdmap e3: 0 total, 0 up, 0 in 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mgrmap e9: x(active, starting, since 0.00692574s) 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mgr metadata", "who": "x", "id": "x"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Manager daemon x is now available 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mgrmap e10: x(active, since 1.00968s) 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: [10/Mar/2026:05:29:10] ENGINE Bus STARTING 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: [10/Mar/2026:05:29:10] ENGINE Client ('192.168.123.101', 45130) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: [10/Mar/2026:05:29:10] ENGINE Serving on https://192.168.123.101:7150 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: [10/Mar/2026:05:29:10] ENGINE Serving on http://192.168.123.101:8765 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: [10/Mar/2026:05:29:10] ENGINE Bus STARTED 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3707339119' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3253245897' entity='client.admin' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mgrmap e11: x(active, since 2s) 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1595755798' entity='client.admin' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Updating vm01:/etc/ceph/ceph.conf 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Updating vm01:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Updating vm01:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.client.admin.keyring 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Deploying cephadm binary to vm02 2026-03-10T05:29:25.311 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Added host vm02 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mgrmap e12: x(active, since 6s) 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm05", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Deploying cephadm binary to vm05 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Added host vm05 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Updating vm02:/etc/ceph/ceph.conf 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Updating vm02:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Updating vm02:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.client.admin.keyring 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1345241876' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1345241876' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm01:192.168.123.101=a;vm02:192.168.123.102=b;vm05:192.168.123.105=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Saving service mon spec with placement vm01:192.168.123.101=a;vm02:192.168.123.102=b;vm05:192.168.123.105=c;count:3 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1762770644' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd/host:vm05", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Updating vm05:/etc/ceph/ceph.conf 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Updating vm05:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Updating vm05:/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Updating vm05:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.client.admin.keyring 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.312 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:25.313 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:29:25.313 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:25.313 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: Deploying daemon mon.c on vm05 2026-03-10T05:29:25.313 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/748299835' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:29:25.313 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:24 vm05 ceph-mon[50927]: mon.c@-1(synchronizing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 3 2026-03-10T05:29:26.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:26 vm02 ceph-mon[50473]: mon.b@-1(synchronizing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 3 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: Deploying daemon mon.b on vm02 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: mon.a calling monitor election 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: mon.c calling monitor election 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: monmap epoch 2 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: last_changed 2026-03-10T05:29:24.890961+0000 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: created 2026-03-10T05:28:47.264782+0000 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: min_mon_release 19 (squid) 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: election_strategy: 1 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.c 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: fsmap 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: mgrmap e12: x(active, since 20s) 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: overall HEALTH_OK 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:30.268 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:29 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:30.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: Deploying daemon mon.b on vm02 2026-03-10T05:29:30.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:29:30.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: mon.a calling monitor election 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: mon.c calling monitor election 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: monmap epoch 2 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: last_changed 2026-03-10T05:29:24.890961+0000 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: created 2026-03-10T05:28:47.264782+0000 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: min_mon_release 19 (squid) 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: election_strategy: 1 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.c 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: fsmap 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: mgrmap e12: x(active, since 20s) 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: overall HEALTH_OK 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:30.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:29 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:31.338 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:30 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:30.887+0000 7fa5c74d4640 -1 mgr.server handle_report got status from non-daemon mon.c 2026-03-10T05:29:35.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: mon.a calling monitor election 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: mon.c calling monitor election 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: mon.b calling monitor election 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: monmap epoch 3 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: last_changed 2026-03-10T05:29:30.211443+0000 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: created 2026-03-10T05:28:47.264782+0000 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: min_mon_release 19 (squid) 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: election_strategy: 1 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.c 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: 2: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.b 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: fsmap 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: mgrmap e12: x(active, since 25s) 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: overall HEALTH_OK 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:35.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:35.567 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:29:35.567 INFO:teuthology.orchestra.run.vm05.stdout:{"epoch":3,"fsid":"dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba","modified":"2026-03-10T05:29:30.211443Z","created":"2026-03-10T05:28:47.264782Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:3300","nonce":0},{"type":"v1","addr":"192.168.123.105:6789","nonce":0}]},"addr":"192.168.123.105:6789/0","public_addr":"192.168.123.105:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":2,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:3300","nonce":0},{"type":"v1","addr":"192.168.123.102:6789","nonce":0}]},"addr":"192.168.123.102:6789/0","public_addr":"192.168.123.102:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1,2]} 2026-03-10T05:29:35.567 INFO:teuthology.orchestra.run.vm05.stderr:dumped monmap epoch 3 2026-03-10T05:29:35.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: mon.a calling monitor election 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: mon.c calling monitor election 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: mon.b calling monitor election 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: monmap epoch 3 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: last_changed 2026-03-10T05:29:30.211443+0000 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: created 2026-03-10T05:28:47.264782+0000 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: min_mon_release 19 (squid) 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: election_strategy: 1 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.c 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: 2: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.b 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: fsmap 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: mgrmap e12: x(active, since 25s) 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: overall HEALTH_OK 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:35.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:35.632 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-10T05:29:35.632 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph config generate-minimal-conf 2026-03-10T05:29:35.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: Deploying daemon mon.b on vm02 2026-03-10T05:29:35.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:29:35.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:35.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: mon.a calling monitor election 2026-03-10T05:29:35.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:35.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: mon.c calling monitor election 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: monmap epoch 2 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: last_changed 2026-03-10T05:29:24.890961+0000 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: created 2026-03-10T05:28:47.264782+0000 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: min_mon_release 19 (squid) 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: election_strategy: 1 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.c 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: fsmap 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: mgrmap e12: x(active, since 20s) 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: overall HEALTH_OK 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: mon.a calling monitor election 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: mon.c calling monitor election 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: mon.b calling monitor election 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: monmap epoch 3 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: last_changed 2026-03-10T05:29:30.211443+0000 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: created 2026-03-10T05:28:47.264782+0000 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: min_mon_release 19 (squid) 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: election_strategy: 1 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: 0: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.a 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: 1: [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] mon.c 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: 2: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.b 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: fsmap 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: mgrmap e12: x(active, since 25s) 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: overall HEALTH_OK 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:35.649 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:35.843 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:29:36.095 INFO:teuthology.orchestra.run.vm01.stdout:# minimal ceph.conf for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:36.095 INFO:teuthology.orchestra.run.vm01.stdout:[global] 2026-03-10T05:29:36.095 INFO:teuthology.orchestra.run.vm01.stdout: fsid = dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:29:36.095 INFO:teuthology.orchestra.run.vm01.stdout: mon_host = [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] [v2:192.168.123.105:3300/0,v1:192.168.123.105:6789/0] 2026-03-10T05:29:36.174 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-10T05:29:36.174 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:29:36.174 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T05:29:36.208 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:29:36.208 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:29:36.283 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:29:36.284 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T05:29:36.314 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:29:36.314 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:29:36.380 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T05:29:36.380 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T05:29:36.416 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T05:29:36.416 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T05:29:36.502 INFO:tasks.cephadm:Adding mgr.x on vm01 2026-03-10T05:29:36.502 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch apply mgr '1;vm01=x' 2026-03-10T05:29:36.529 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: Updating vm01:/etc/ceph/ceph.conf 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: Updating vm01:/etc/ceph/ceph.conf 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: Updating vm02:/etc/ceph/ceph.conf 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: Updating vm05:/etc/ceph/ceph.conf 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: Updating vm05:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: Updating vm02:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: Updating vm01:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/2141626069' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.530 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.531 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:29:36.531 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T05:29:36.531 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:36.531 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3396152199' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:36.531 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:36.531 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.531 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.531 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:29:36.531 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T05:29:36.531 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:36.531 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.531 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.531 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:36.716 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.c/config 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: Updating vm01:/etc/ceph/ceph.conf 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: Updating vm02:/etc/ceph/ceph.conf 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: Updating vm05:/etc/ceph/ceph.conf 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: Updating vm05:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: Updating vm02:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: Updating vm01:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/2141626069' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3396152199' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.743 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: Updating vm02:/etc/ceph/ceph.conf 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: Updating vm05:/etc/ceph/ceph.conf 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: Updating vm05:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: Updating vm02:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: Updating vm01:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/config/ceph.conf 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/2141626069' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3396152199' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:36.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:36.969 INFO:teuthology.orchestra.run.vm05.stdout:Scheduled mgr update... 2026-03-10T05:29:37.024 INFO:tasks.cephadm:Deploying OSDs... 2026-03-10T05:29:37.024 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:29:37.024 DEBUG:teuthology.orchestra.run.vm01:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T05:29:37.046 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:29:37.047 DEBUG:teuthology.orchestra.run.vm01:> ls /dev/[sv]d? 2026-03-10T05:29:37.108 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vda 2026-03-10T05:29:37.108 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdb 2026-03-10T05:29:37.108 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdc 2026-03-10T05:29:37.108 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdd 2026-03-10T05:29:37.108 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vde 2026-03-10T05:29:37.109 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T05:29:37.109 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T05:29:37.109 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdb 2026-03-10T05:29:37.169 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdb 2026-03-10T05:29:37.169 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:29:37.169 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-10T05:29:37.169 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:29:37.169 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:29:37.169 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 05:29:13.231538843 +0000 2026-03-10T05:29:37.169 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 05:24:16.000000000 +0000 2026-03-10T05:29:37.169 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 05:24:16.000000000 +0000 2026-03-10T05:29:37.169 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-10 05:24:14.191000000 +0000 2026-03-10T05:29:37.169 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T05:29:37.261 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T05:29:37.261 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T05:29:37.261 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 9.4308e-05 s, 5.4 MB/s 2026-03-10T05:29:37.263 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T05:29:37.299 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdc 2026-03-10T05:29:37.357 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdc 2026-03-10T05:29:37.358 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:29:37.358 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-10T05:29:37.358 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:29:37.358 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:29:37.358 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 05:29:13.282538899 +0000 2026-03-10T05:29:37.358 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 05:24:16.014000000 +0000 2026-03-10T05:29:37.358 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 05:24:16.014000000 +0000 2026-03-10T05:29:37.358 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-10 05:24:14.202000000 +0000 2026-03-10T05:29:37.358 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T05:29:37.428 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T05:29:37.428 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T05:29:37.428 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000193952 s, 2.6 MB/s 2026-03-10T05:29:37.429 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T05:29:37.489 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdd 2026-03-10T05:29:37.534 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: Reconfiguring mon.a (unknown last config time)... 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: Reconfiguring daemon mon.a on vm01 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: Reconfiguring mon.b (monmap changed)... 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: Reconfiguring daemon mon.b on vm02 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: Reconfiguring mon.c (monmap changed)... 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: Reconfiguring daemon mon.c on vm05 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:37 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.535 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:29:37 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x[48149]: 2026-03-10T05:29:37.210+0000 7fa5c74d4640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-10T05:29:37.552 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdd 2026-03-10T05:29:37.552 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:29:37.552 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-10T05:29:37.552 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:29:37.552 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:29:37.552 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 05:29:13.332538955 +0000 2026-03-10T05:29:37.552 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 05:24:16.072000000 +0000 2026-03-10T05:29:37.552 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 05:24:16.072000000 +0000 2026-03-10T05:29:37.552 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-10 05:24:14.209000000 +0000 2026-03-10T05:29:37.552 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T05:29:37.619 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T05:29:37.619 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T05:29:37.619 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000179185 s, 2.9 MB/s 2026-03-10T05:29:37.620 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T05:29:37.677 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vde 2026-03-10T05:29:37.736 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vde 2026-03-10T05:29:37.736 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:29:37.736 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-10T05:29:37.736 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:29:37.736 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:29:37.736 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-10 05:29:13.390539020 +0000 2026-03-10T05:29:37.736 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-10 05:24:16.002000000 +0000 2026-03-10T05:29:37.736 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-10 05:24:16.002000000 +0000 2026-03-10T05:29:37.736 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-10 05:24:14.215000000 +0000 2026-03-10T05:29:37.736 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T05:29:37.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: Reconfiguring mon.a (unknown last config time)... 2026-03-10T05:29:37.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: Reconfiguring daemon mon.a on vm01 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: Reconfiguring mon.b (monmap changed)... 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: Reconfiguring daemon mon.b on vm02 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: Reconfiguring mon.c (monmap changed)... 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: Reconfiguring daemon mon.c on vm05 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:37 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.806 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-10T05:29:37.806 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-10T05:29:37.806 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.00015419 s, 3.3 MB/s 2026-03-10T05:29:37.807 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T05:29:37.871 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:29:37.871 DEBUG:teuthology.orchestra.run.vm02:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T05:29:37.887 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:29:37.887 DEBUG:teuthology.orchestra.run.vm02:> ls /dev/[sv]d? 2026-03-10T05:29:37.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: Reconfiguring mon.a (unknown last config time)... 2026-03-10T05:29:37.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: Reconfiguring daemon mon.a on vm01 2026-03-10T05:29:37.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:37.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: Reconfiguring mon.b (monmap changed)... 2026-03-10T05:29:37.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: Reconfiguring daemon mon.b on vm02 2026-03-10T05:29:37.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: Reconfiguring mon.c (monmap changed)... 2026-03-10T05:29:37.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: Reconfiguring daemon mon.c on vm05 2026-03-10T05:29:37.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:37.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:37.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:37.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:37.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:37.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T05:29:37.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T05:29:37.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:37.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:37 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:37.914 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vda 2026-03-10T05:29:37.914 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdb 2026-03-10T05:29:37.914 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdc 2026-03-10T05:29:37.914 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdd 2026-03-10T05:29:37.914 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vde 2026-03-10T05:29:37.914 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T05:29:37.914 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T05:29:37.914 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdb 2026-03-10T05:29:37.972 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdb 2026-03-10T05:29:37.972 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:29:37.972 INFO:teuthology.orchestra.run.vm02.stdout:Device: 6h/6d Inode: 219 Links: 1 Device type: fc,10 2026-03-10T05:29:37.972 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:29:37.972 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:29:37.972 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-10 05:29:18.781721308 +0000 2026-03-10T05:29:37.972 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-10 05:24:46.998000000 +0000 2026-03-10T05:29:37.972 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-10 05:24:46.998000000 +0000 2026-03-10T05:29:37.972 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-03-10 05:24:45.232000000 +0000 2026-03-10T05:29:37.972 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T05:29:38.034 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-10T05:29:38.034 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-10T05:29:38.034 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.00014976 s, 3.4 MB/s 2026-03-10T05:29:38.035 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T05:29:38.092 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdc 2026-03-10T05:29:38.148 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdc 2026-03-10T05:29:38.148 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:29:38.148 INFO:teuthology.orchestra.run.vm02.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-10T05:29:38.148 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:29:38.148 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:29:38.148 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-10 05:29:18.806721308 +0000 2026-03-10T05:29:38.148 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-10 05:24:47.026000000 +0000 2026-03-10T05:29:38.148 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-10 05:24:47.026000000 +0000 2026-03-10T05:29:38.148 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-03-10 05:24:45.253000000 +0000 2026-03-10T05:29:38.148 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T05:29:38.213 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-10T05:29:38.213 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-10T05:29:38.213 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000193943 s, 2.6 MB/s 2026-03-10T05:29:38.214 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T05:29:38.272 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdd 2026-03-10T05:29:38.328 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdd 2026-03-10T05:29:38.328 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:29:38.328 INFO:teuthology.orchestra.run.vm02.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-10T05:29:38.328 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:29:38.328 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:29:38.328 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-10 05:29:18.833721308 +0000 2026-03-10T05:29:38.328 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-10 05:24:47.004000000 +0000 2026-03-10T05:29:38.328 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-10 05:24:47.004000000 +0000 2026-03-10T05:29:38.328 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-03-10 05:24:45.260000000 +0000 2026-03-10T05:29:38.328 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T05:29:38.391 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-10T05:29:38.391 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-10T05:29:38.391 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000150321 s, 3.4 MB/s 2026-03-10T05:29:38.392 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T05:29:38.448 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vde 2026-03-10T05:29:38.505 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vde 2026-03-10T05:29:38.505 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:29:38.505 INFO:teuthology.orchestra.run.vm02.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-10T05:29:38.505 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:29:38.505 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:29:38.505 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-10 05:29:18.864721307 +0000 2026-03-10T05:29:38.505 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-10 05:24:47.062000000 +0000 2026-03-10T05:29:38.505 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-10 05:24:47.062000000 +0000 2026-03-10T05:29:38.505 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-03-10 05:24:45.265000000 +0000 2026-03-10T05:29:38.506 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T05:29:38.570 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-10T05:29:38.570 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-10T05:29:38.570 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000158747 s, 3.2 MB/s 2026-03-10T05:29:38.571 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T05:29:38.630 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T05:29:38.630 DEBUG:teuthology.orchestra.run.vm05:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T05:29:38.646 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:29:38.646 DEBUG:teuthology.orchestra.run.vm05:> ls /dev/[sv]d? 2026-03-10T05:29:38.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:38 vm02 ceph-mon[50473]: from='client.24103 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "1;vm01=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:38.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:38 vm02 ceph-mon[50473]: Saving service mgr spec with placement vm01=x;count:1 2026-03-10T05:29:38.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:38 vm02 ceph-mon[50473]: Reconfiguring mgr.x (unknown last config time)... 2026-03-10T05:29:38.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:38 vm02 ceph-mon[50473]: Reconfiguring daemon mgr.x on vm01 2026-03-10T05:29:38.705 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vda 2026-03-10T05:29:38.705 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vdb 2026-03-10T05:29:38.705 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vdc 2026-03-10T05:29:38.706 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vdd 2026-03-10T05:29:38.706 INFO:teuthology.orchestra.run.vm05.stdout:/dev/vde 2026-03-10T05:29:38.706 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T05:29:38.706 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T05:29:38.706 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vdb 2026-03-10T05:29:38.764 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vdb 2026-03-10T05:29:38.765 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:29:38.765 INFO:teuthology.orchestra.run.vm05.stdout:Device: 6h/6d Inode: 223 Links: 1 Device type: fc,10 2026-03-10T05:29:38.765 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:29:38.765 INFO:teuthology.orchestra.run.vm05.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:29:38.765 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-10 05:29:22.496201867 +0000 2026-03-10T05:29:38.765 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-10 05:25:11.950000000 +0000 2026-03-10T05:29:38.765 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-10 05:25:11.950000000 +0000 2026-03-10T05:29:38.765 INFO:teuthology.orchestra.run.vm05.stdout: Birth: 2026-03-10 05:25:10.181000000 +0000 2026-03-10T05:29:38.765 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T05:29:38.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:38 vm05 ceph-mon[50927]: from='client.24103 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "1;vm01=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:38.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:38 vm05 ceph-mon[50927]: Saving service mgr spec with placement vm01=x;count:1 2026-03-10T05:29:38.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:38 vm05 ceph-mon[50927]: Reconfiguring mgr.x (unknown last config time)... 2026-03-10T05:29:38.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:38 vm05 ceph-mon[50927]: Reconfiguring daemon mgr.x on vm01 2026-03-10T05:29:38.832 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-10T05:29:38.833 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-10T05:29:38.833 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000178644 s, 2.9 MB/s 2026-03-10T05:29:38.834 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T05:29:38.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:38 vm01 ceph-mon[47941]: from='client.24103 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "1;vm01=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:38.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:38 vm01 ceph-mon[47941]: Saving service mgr spec with placement vm01=x;count:1 2026-03-10T05:29:38.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:38 vm01 ceph-mon[47941]: Reconfiguring mgr.x (unknown last config time)... 2026-03-10T05:29:38.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:38 vm01 ceph-mon[47941]: Reconfiguring daemon mgr.x on vm01 2026-03-10T05:29:38.891 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vdc 2026-03-10T05:29:38.951 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vdc 2026-03-10T05:29:38.951 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:29:38.951 INFO:teuthology.orchestra.run.vm05.stdout:Device: 6h/6d Inode: 224 Links: 1 Device type: fc,20 2026-03-10T05:29:38.951 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:29:38.951 INFO:teuthology.orchestra.run.vm05.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:29:38.951 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-10 05:29:22.527201880 +0000 2026-03-10T05:29:38.951 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-10 05:25:11.962000000 +0000 2026-03-10T05:29:38.951 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-10 05:25:11.962000000 +0000 2026-03-10T05:29:38.951 INFO:teuthology.orchestra.run.vm05.stdout: Birth: 2026-03-10 05:25:10.191000000 +0000 2026-03-10T05:29:38.951 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T05:29:39.016 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-10T05:29:39.016 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-10T05:29:39.016 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000241853 s, 2.1 MB/s 2026-03-10T05:29:39.017 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T05:29:39.077 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vdd 2026-03-10T05:29:39.138 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vdd 2026-03-10T05:29:39.138 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:29:39.138 INFO:teuthology.orchestra.run.vm05.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-10T05:29:39.138 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:29:39.138 INFO:teuthology.orchestra.run.vm05.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:29:39.138 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-10 05:29:22.552201890 +0000 2026-03-10T05:29:39.138 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-10 05:25:11.963000000 +0000 2026-03-10T05:29:39.138 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-10 05:25:11.963000000 +0000 2026-03-10T05:29:39.138 INFO:teuthology.orchestra.run.vm05.stdout: Birth: 2026-03-10 05:25:10.202000000 +0000 2026-03-10T05:29:39.138 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T05:29:39.208 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-10T05:29:39.208 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-10T05:29:39.208 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.0001767 s, 2.9 MB/s 2026-03-10T05:29:39.209 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T05:29:39.266 DEBUG:teuthology.orchestra.run.vm05:> stat /dev/vde 2026-03-10T05:29:39.323 INFO:teuthology.orchestra.run.vm05.stdout: File: /dev/vde 2026-03-10T05:29:39.323 INFO:teuthology.orchestra.run.vm05.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T05:29:39.323 INFO:teuthology.orchestra.run.vm05.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-10T05:29:39.323 INFO:teuthology.orchestra.run.vm05.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T05:29:39.323 INFO:teuthology.orchestra.run.vm05.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T05:29:39.323 INFO:teuthology.orchestra.run.vm05.stdout:Access: 2026-03-10 05:29:22.577201901 +0000 2026-03-10T05:29:39.323 INFO:teuthology.orchestra.run.vm05.stdout:Modify: 2026-03-10 05:25:11.952000000 +0000 2026-03-10T05:29:39.323 INFO:teuthology.orchestra.run.vm05.stdout:Change: 2026-03-10 05:25:11.952000000 +0000 2026-03-10T05:29:39.323 INFO:teuthology.orchestra.run.vm05.stdout: Birth: 2026-03-10 05:25:10.205000000 +0000 2026-03-10T05:29:39.324 DEBUG:teuthology.orchestra.run.vm05:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T05:29:39.388 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records in 2026-03-10T05:29:39.388 INFO:teuthology.orchestra.run.vm05.stderr:1+0 records out 2026-03-10T05:29:39.388 INFO:teuthology.orchestra.run.vm05.stderr:512 bytes copied, 0.000200005 s, 2.6 MB/s 2026-03-10T05:29:39.389 DEBUG:teuthology.orchestra.run.vm05:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T05:29:39.448 INFO:tasks.cephadm:Deploying osd.0 on vm01 with /dev/vde... 2026-03-10T05:29:39.449 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- lvm zap /dev/vde 2026-03-10T05:29:39.540 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:39 vm05 ceph-mon[50927]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:39.624 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:29:39.835 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:39 vm01 ceph-mon[47941]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:39.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:39 vm02 ceph-mon[50473]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:40.519 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:29:40.540 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch daemon add osd vm01:/dev/vde 2026-03-10T05:29:40.710 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:29:40.734 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:40 vm01 ceph-mon[47941]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:40.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:40 vm05 ceph-mon[50927]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:40.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:40 vm02 ceph-mon[50473]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:41.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:41 vm01 ceph-mon[47941]: from='client.14211 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:41.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:41 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:29:41.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:41 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:29:41.718 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:41 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:42.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:41 vm05 ceph-mon[50927]: from='client.14211 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:42.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:41 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:29:42.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:41 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:29:42.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:41 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:42.147 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:41 vm02 ceph-mon[50473]: from='client.14211 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:42.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:41 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:29:42.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:41 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:29:42.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:41 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:42.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:42 vm02 ceph-mon[50473]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:42.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:42 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3309622891' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e9a77b3b-eb3e-404d-816d-a374f1e03f47"}]: dispatch 2026-03-10T05:29:42.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:42 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3309622891' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e9a77b3b-eb3e-404d-816d-a374f1e03f47"}]': finished 2026-03-10T05:29:42.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:42 vm02 ceph-mon[50473]: osdmap e5: 1 total, 0 up, 1 in 2026-03-10T05:29:42.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:42 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:42.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:42 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2454411440' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:29:43.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:42 vm05 ceph-mon[50927]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:43.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:42 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3309622891' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e9a77b3b-eb3e-404d-816d-a374f1e03f47"}]: dispatch 2026-03-10T05:29:43.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:42 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3309622891' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e9a77b3b-eb3e-404d-816d-a374f1e03f47"}]': finished 2026-03-10T05:29:43.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:42 vm05 ceph-mon[50927]: osdmap e5: 1 total, 0 up, 1 in 2026-03-10T05:29:43.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:42 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:43.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:42 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2454411440' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:29:43.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:42 vm01 ceph-mon[47941]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:43.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:42 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3309622891' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e9a77b3b-eb3e-404d-816d-a374f1e03f47"}]: dispatch 2026-03-10T05:29:43.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:42 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3309622891' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e9a77b3b-eb3e-404d-816d-a374f1e03f47"}]': finished 2026-03-10T05:29:43.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:42 vm01 ceph-mon[47941]: osdmap e5: 1 total, 0 up, 1 in 2026-03-10T05:29:43.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:42 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:43.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:42 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2454411440' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:29:45.085 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:44 vm01 ceph-mon[47941]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:45.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:44 vm05 ceph-mon[50927]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:45.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:44 vm02 ceph-mon[50473]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:46.035 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:45 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T05:29:46.035 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:45 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:46.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:45 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T05:29:46.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:45 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:46.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:45 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T05:29:46.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:45 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:47.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:46 vm05 ceph-mon[50927]: Deploying daemon osd.0 on vm01 2026-03-10T05:29:47.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:46 vm05 ceph-mon[50927]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:47.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:46 vm01 ceph-mon[47941]: Deploying daemon osd.0 on vm01 2026-03-10T05:29:47.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:46 vm01 ceph-mon[47941]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:47.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:46 vm02 ceph-mon[50473]: Deploying daemon osd.0 on vm01 2026-03-10T05:29:47.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:46 vm02 ceph-mon[50473]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:49.067 INFO:teuthology.orchestra.run.vm01.stdout:Created osd(s) 0 on host 'vm01' 2026-03-10T05:29:49.126 DEBUG:teuthology.orchestra.run.vm01:osd.0> sudo journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.0.service 2026-03-10T05:29:49.128 INFO:tasks.cephadm:Deploying osd.1 on vm01 with /dev/vdd... 2026-03-10T05:29:49.128 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- lvm zap /dev/vdd 2026-03-10T05:29:49.202 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:48 vm01 ceph-mon[47941]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:49.202 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:48 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:49.202 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:48 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.202 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:48 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.202 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:48 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.202 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:48 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.202 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:48 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:49.202 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:48 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:49.202 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:48 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:48 vm05 ceph-mon[50927]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:49.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:48 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:49.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:48 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:48 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:48 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:48 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:48 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:49.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:48 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:49.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:48 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:48 vm02 ceph-mon[50473]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:48 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:48 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:48 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:48 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:48 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:48 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:48 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:48 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:49.440 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:29:49.476 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:29:49 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-0[57438]: 2026-03-10T05:29:49.293+0000 7f8b56588740 -1 Falling back to public interface 2026-03-10T05:29:50.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:49 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:50.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:49 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:50.215 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:49 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:50.215 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:29:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-0[57438]: 2026-03-10T05:29:50.128+0000 7f8b56588740 -1 osd.0 0 log_to_monitors true 2026-03-10T05:29:50.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:49 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:50.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:49 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:50.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:49 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:50.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:49 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:29:50.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:49 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:50.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:49 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:50.889 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:29:50.911 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch daemon add osd vm01:/dev/vdd 2026-03-10T05:29:51.142 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:29:51.503 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:51 vm01 ceph-mon[47941]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:51.503 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:51 vm01 ceph-mon[47941]: Detected new or changed devices on vm01 2026-03-10T05:29:51.503 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:51 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:51.503 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:51 vm01 ceph-mon[47941]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T05:29:51.503 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:51 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:51.503 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:51 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:29:51.503 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:51 vm01 ceph-mon[47941]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-10T05:29:51.503 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:51 vm01 ceph-mon[47941]: Unable to set osd_memory_target on vm01 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-10T05:29:51.503 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:51 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:51.503 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:51 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:51.503 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:51 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:51.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:51 vm05 ceph-mon[50927]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:51.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:51 vm05 ceph-mon[50927]: Detected new or changed devices on vm01 2026-03-10T05:29:51.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:51 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:51.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:51 vm05 ceph-mon[50927]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T05:29:51.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:51 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:51.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:51 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:29:51.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:51 vm05 ceph-mon[50927]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-10T05:29:51.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:51 vm05 ceph-mon[50927]: Unable to set osd_memory_target on vm01 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-10T05:29:51.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:51 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:51.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:51 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:51.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:51 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:51.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:51 vm02 ceph-mon[50473]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:51.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:51 vm02 ceph-mon[50473]: Detected new or changed devices on vm01 2026-03-10T05:29:51.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:51 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:51.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:51 vm02 ceph-mon[50473]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T05:29:51.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:51 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:51.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:51 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:29:51.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:51 vm02 ceph-mon[50473]: Adjusting osd_memory_target on vm01 to 257.0M 2026-03-10T05:29:51.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:51 vm02 ceph-mon[50473]: Unable to set osd_memory_target on vm01 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-10T05:29:51.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:51 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:51.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:51 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:29:51.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:51 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:29:52.498 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:52 vm01 ceph-mon[47941]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T05:29:52.499 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:52 vm01 ceph-mon[47941]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T05:29:52.499 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:52 vm01 ceph-mon[47941]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:29:52.499 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:52 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:52.499 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:52 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:29:52.499 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:52 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:29:52.499 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:52 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:52.499 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:52 vm01 ceph-mon[47941]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T05:29:52.499 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:52 vm01 ceph-mon[47941]: osdmap e7: 1 total, 0 up, 1 in 2026-03-10T05:29:52.499 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:52 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:52.499 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:52 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:52.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:52 vm05 ceph-mon[50927]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T05:29:52.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:52 vm05 ceph-mon[50927]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T05:29:52.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:52 vm05 ceph-mon[50927]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:29:52.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:52 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:52.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:52 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:29:52.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:52 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:29:52.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:52 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:52.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:52 vm05 ceph-mon[50927]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T05:29:52.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:52 vm05 ceph-mon[50927]: osdmap e7: 1 total, 0 up, 1 in 2026-03-10T05:29:52.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:52 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:52.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:52 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:52.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:52 vm02 ceph-mon[50473]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T05:29:52.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:52 vm02 ceph-mon[50473]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T05:29:52.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:52 vm02 ceph-mon[50473]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:29:52.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:52 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:52.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:52 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:29:52.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:52 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:29:52.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:52 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:52.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:52 vm02 ceph-mon[50473]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T05:29:52.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:52 vm02 ceph-mon[50473]: osdmap e7: 1 total, 0 up, 1 in 2026-03-10T05:29:52.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:52 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:52.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:52 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:53.754 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:53 vm01 ceph-mon[47941]: purged_snaps scrub starts 2026-03-10T05:29:53.754 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:53 vm01 ceph-mon[47941]: purged_snaps scrub ok 2026-03-10T05:29:53.754 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:53 vm01 ceph-mon[47941]: from='client.14238 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:53.754 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:53 vm01 ceph-mon[47941]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:53.754 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:53 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3030551228' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "423e15ff-ce61-4730-b91e-0f9ca81dd61d"}]: dispatch 2026-03-10T05:29:53.754 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:53 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3030551228' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "423e15ff-ce61-4730-b91e-0f9ca81dd61d"}]': finished 2026-03-10T05:29:53.754 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:53 vm01 ceph-mon[47941]: osdmap e8: 2 total, 0 up, 2 in 2026-03-10T05:29:53.754 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:53 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:53.754 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:53 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:29:53.754 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:53 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2106444022' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:29:53.754 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:53 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:53 vm05 ceph-mon[50927]: purged_snaps scrub starts 2026-03-10T05:29:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:53 vm05 ceph-mon[50927]: purged_snaps scrub ok 2026-03-10T05:29:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:53 vm05 ceph-mon[50927]: from='client.14238 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:53 vm05 ceph-mon[50927]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:53 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3030551228' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "423e15ff-ce61-4730-b91e-0f9ca81dd61d"}]: dispatch 2026-03-10T05:29:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:53 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3030551228' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "423e15ff-ce61-4730-b91e-0f9ca81dd61d"}]': finished 2026-03-10T05:29:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:53 vm05 ceph-mon[50927]: osdmap e8: 2 total, 0 up, 2 in 2026-03-10T05:29:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:53 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:53 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:29:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:53 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2106444022' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:29:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:53 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:53.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:53 vm02 ceph-mon[50473]: purged_snaps scrub starts 2026-03-10T05:29:53.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:53 vm02 ceph-mon[50473]: purged_snaps scrub ok 2026-03-10T05:29:53.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:53 vm02 ceph-mon[50473]: from='client.14238 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:29:53.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:53 vm02 ceph-mon[50473]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:53.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:53 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3030551228' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "423e15ff-ce61-4730-b91e-0f9ca81dd61d"}]: dispatch 2026-03-10T05:29:53.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:53 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3030551228' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "423e15ff-ce61-4730-b91e-0f9ca81dd61d"}]': finished 2026-03-10T05:29:53.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:53 vm02 ceph-mon[50473]: osdmap e8: 2 total, 0 up, 2 in 2026-03-10T05:29:53.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:53 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:53.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:53 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:29:53.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:53 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2106444022' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:29:53.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:53 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:54.088 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:29:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-0[57438]: 2026-03-10T05:29:53.749+0000 7f8b52509640 -1 osd.0 0 waiting for initial osdmap 2026-03-10T05:29:54.088 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:29:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-0[57438]: 2026-03-10T05:29:53.757+0000 7f8b4db32640 -1 osd.0 8 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T05:29:55.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:54 vm05 ceph-mon[50927]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' 2026-03-10T05:29:55.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:54 vm05 ceph-mon[50927]: pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:55.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:54 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:54 vm01 ceph-mon[47941]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' 2026-03-10T05:29:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:54 vm01 ceph-mon[47941]: pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:54 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:55.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:54 vm02 ceph-mon[50473]: from='osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884]' entity='osd.0' 2026-03-10T05:29:55.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:54 vm02 ceph-mon[50473]: pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T05:29:55.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:54 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:56.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:55 vm01 ceph-mon[47941]: osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884] boot 2026-03-10T05:29:56.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:55 vm01 ceph-mon[47941]: osdmap e9: 2 total, 1 up, 2 in 2026-03-10T05:29:56.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:55 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:56.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:55 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:29:56.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:55 vm05 ceph-mon[50927]: osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884] boot 2026-03-10T05:29:56.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:55 vm05 ceph-mon[50927]: osdmap e9: 2 total, 1 up, 2 in 2026-03-10T05:29:56.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:55 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:56.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:55 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:29:56.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:55 vm02 ceph-mon[50473]: osd.0 [v2:192.168.123.101:6802/2152256884,v1:192.168.123.101:6803/2152256884] boot 2026-03-10T05:29:56.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:55 vm02 ceph-mon[50473]: osdmap e9: 2 total, 1 up, 2 in 2026-03-10T05:29:56.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:55 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T05:29:56.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:55 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:29:57.040 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:56 vm01 ceph-mon[47941]: pgmap v22: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:29:57.040 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:56 vm01 ceph-mon[47941]: osdmap e10: 2 total, 1 up, 2 in 2026-03-10T05:29:57.040 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:56 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:29:57.040 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:56 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T05:29:57.040 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:56 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:57.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:56 vm05 ceph-mon[50927]: pgmap v22: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:29:57.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:56 vm05 ceph-mon[50927]: osdmap e10: 2 total, 1 up, 2 in 2026-03-10T05:29:57.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:56 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:29:57.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:56 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T05:29:57.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:56 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:57.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:56 vm02 ceph-mon[50473]: pgmap v22: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:29:57.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:56 vm02 ceph-mon[50473]: osdmap e10: 2 total, 1 up, 2 in 2026-03-10T05:29:57.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:56 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:29:57.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:56 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T05:29:57.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:56 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:29:58.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:57 vm05 ceph-mon[50927]: Deploying daemon osd.1 on vm01 2026-03-10T05:29:58.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:57 vm01 ceph-mon[47941]: Deploying daemon osd.1 on vm01 2026-03-10T05:29:58.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:57 vm02 ceph-mon[50473]: Deploying daemon osd.1 on vm01 2026-03-10T05:29:59.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:58 vm05 ceph-mon[50927]: pgmap v24: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:29:59.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:58 vm01 ceph-mon[47941]: pgmap v24: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:29:59.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:58 vm02 ceph-mon[50473]: pgmap v24: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:29:59.965 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:59 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:00.208 INFO:teuthology.orchestra.run.vm01.stdout:Created osd(s) 1 on host 'vm01' 2026-03-10T05:30:00.226 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:59 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:59 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:59 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:59 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:59 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:00.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:59 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:00.227 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:29:59 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.275 DEBUG:teuthology.orchestra.run.vm01:osd.1> sudo journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.1.service 2026-03-10T05:30:00.277 INFO:tasks.cephadm:Deploying osd.2 on vm02 with /dev/vde... 2026-03-10T05:30:00.277 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- lvm zap /dev/vde 2026-03-10T05:30:00.298 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:59 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:00.298 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:59 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.298 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:59 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.298 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:59 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.298 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:59 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.298 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:59 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:00.298 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:59 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:00.298 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:29:59 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:59 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:59 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:59 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:59 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:59 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:59 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:59 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:29:59 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:00.452 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.b/config 2026-03-10T05:30:00.739 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:30:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-1[62386]: 2026-03-10T05:30:00.733+0000 7f3906ef8740 -1 osd.1 0 log_to_monitors true 2026-03-10T05:30:01.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:00 vm01 ceph-mon[47941]: pgmap v25: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:30:01.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:00 vm01 ceph-mon[47941]: overall HEALTH_OK 2026-03-10T05:30:01.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:00 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:01.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:00 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:01.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:00 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:01.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:00 vm01 ceph-mon[47941]: from='osd.1 [v2:192.168.123.101:6810/1960293858,v1:192.168.123.101:6811/1960293858]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T05:30:01.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:00 vm01 ceph-mon[47941]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T05:30:01.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:00 vm02 ceph-mon[50473]: pgmap v25: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:30:01.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:00 vm02 ceph-mon[50473]: overall HEALTH_OK 2026-03-10T05:30:01.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:00 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:01.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:00 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:01.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:00 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:01.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:00 vm02 ceph-mon[50473]: from='osd.1 [v2:192.168.123.101:6810/1960293858,v1:192.168.123.101:6811/1960293858]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T05:30:01.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:00 vm02 ceph-mon[50473]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T05:30:01.291 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:30:01.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:00 vm05 ceph-mon[50927]: pgmap v25: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:30:01.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:00 vm05 ceph-mon[50927]: overall HEALTH_OK 2026-03-10T05:30:01.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:00 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:01.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:00 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:01.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:00 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:01.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:00 vm05 ceph-mon[50927]: from='osd.1 [v2:192.168.123.101:6810/1960293858,v1:192.168.123.101:6811/1960293858]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T05:30:01.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:00 vm05 ceph-mon[50927]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T05:30:01.311 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch daemon add osd vm02:/dev/vde 2026-03-10T05:30:01.526 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.b/config 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: Detected new or changed devices on vm01 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: Adjusting osd_memory_target on vm01 to 128.5M 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: Unable to set osd_memory_target on vm01 to 134768230: error parsing value: Value '134768230' is below minimum 939524096 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: osdmap e11: 2 total, 1 up, 2 in 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: from='osd.1 [v2:192.168.123.101:6810/1960293858,v1:192.168.123.101:6811/1960293858]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:02.381 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:02 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: Detected new or changed devices on vm01 2026-03-10T05:30:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: Adjusting osd_memory_target on vm01 to 128.5M 2026-03-10T05:30:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: Unable to set osd_memory_target on vm01 to 134768230: error parsing value: Value '134768230' is below minimum 939524096 2026-03-10T05:30:02.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T05:30:02.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: osdmap e11: 2 total, 1 up, 2 in 2026-03-10T05:30:02.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:02.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:02.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:30:02.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: from='osd.1 [v2:192.168.123.101:6810/1960293858,v1:192.168.123.101:6811/1960293858]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:30:02.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:02.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:02.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:02.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:02.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:02 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:02.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: Detected new or changed devices on vm01 2026-03-10T05:30:02.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: Adjusting osd_memory_target on vm01 to 128.5M 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: Unable to set osd_memory_target on vm01 to 134768230: error parsing value: Value '134768230' is below minimum 939524096 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: osdmap e11: 2 total, 1 up, 2 in 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: from='osd.1 [v2:192.168.123.101:6810/1960293858,v1:192.168.123.101:6811/1960293858]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:02.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:02 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:03.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:03 vm02 ceph-mon[50473]: from='client.24142 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:03.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:03 vm02 ceph-mon[50473]: pgmap v27: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:30:03.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:03 vm02 ceph-mon[50473]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T05:30:03.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:03 vm02 ceph-mon[50473]: osdmap e12: 2 total, 1 up, 2 in 2026-03-10T05:30:03.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:03 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:03.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:03 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:03.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:03 vm02 ceph-mon[50473]: from='client.? 192.168.123.102:0/2205659159' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e19eb7fe-8764-496e-ad96-eed37a0f93ad"}]: dispatch 2026-03-10T05:30:03.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:03 vm02 ceph-mon[50473]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e19eb7fe-8764-496e-ad96-eed37a0f93ad"}]: dispatch 2026-03-10T05:30:03.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:03 vm02 ceph-mon[50473]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e19eb7fe-8764-496e-ad96-eed37a0f93ad"}]': finished 2026-03-10T05:30:03.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:03 vm02 ceph-mon[50473]: osdmap e13: 3 total, 1 up, 3 in 2026-03-10T05:30:03.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:03 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:03.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:03 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:03.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:03 vm02 ceph-mon[50473]: from='client.? 192.168.123.102:0/3079689825' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:03.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:03 vm02 ceph-mon[50473]: from='osd.1 ' entity='osd.1' 2026-03-10T05:30:03.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:03 vm05 ceph-mon[50927]: from='client.24142 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:03.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:03 vm05 ceph-mon[50927]: pgmap v27: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:30:03.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:03 vm05 ceph-mon[50927]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T05:30:03.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:03 vm05 ceph-mon[50927]: osdmap e12: 2 total, 1 up, 2 in 2026-03-10T05:30:03.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:03 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:03.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:03 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:03.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:03 vm05 ceph-mon[50927]: from='client.? 192.168.123.102:0/2205659159' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e19eb7fe-8764-496e-ad96-eed37a0f93ad"}]: dispatch 2026-03-10T05:30:03.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:03 vm05 ceph-mon[50927]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e19eb7fe-8764-496e-ad96-eed37a0f93ad"}]: dispatch 2026-03-10T05:30:03.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:03 vm05 ceph-mon[50927]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e19eb7fe-8764-496e-ad96-eed37a0f93ad"}]': finished 2026-03-10T05:30:03.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:03 vm05 ceph-mon[50927]: osdmap e13: 3 total, 1 up, 3 in 2026-03-10T05:30:03.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:03 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:03.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:03 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:03.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:03 vm05 ceph-mon[50927]: from='client.? 192.168.123.102:0/3079689825' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:03.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:03 vm05 ceph-mon[50927]: from='osd.1 ' entity='osd.1' 2026-03-10T05:30:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:03 vm01 ceph-mon[47941]: from='client.24142 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:03 vm01 ceph-mon[47941]: pgmap v27: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:30:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:03 vm01 ceph-mon[47941]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-10T05:30:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:03 vm01 ceph-mon[47941]: osdmap e12: 2 total, 1 up, 2 in 2026-03-10T05:30:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:03 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:03 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:03 vm01 ceph-mon[47941]: from='client.? 192.168.123.102:0/2205659159' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e19eb7fe-8764-496e-ad96-eed37a0f93ad"}]: dispatch 2026-03-10T05:30:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:03 vm01 ceph-mon[47941]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e19eb7fe-8764-496e-ad96-eed37a0f93ad"}]: dispatch 2026-03-10T05:30:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:03 vm01 ceph-mon[47941]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e19eb7fe-8764-496e-ad96-eed37a0f93ad"}]': finished 2026-03-10T05:30:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:03 vm01 ceph-mon[47941]: osdmap e13: 3 total, 1 up, 3 in 2026-03-10T05:30:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:03 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:03 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:03 vm01 ceph-mon[47941]: from='client.? 192.168.123.102:0/3079689825' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:03 vm01 ceph-mon[47941]: from='osd.1 ' entity='osd.1' 2026-03-10T05:30:03.589 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:30:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-1[62386]: 2026-03-10T05:30:03.159+0000 7f390368c640 -1 osd.1 0 waiting for initial osdmap 2026-03-10T05:30:03.589 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:30:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-1[62386]: 2026-03-10T05:30:03.166+0000 7f38fe4a2640 -1 osd.1 13 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T05:30:04.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:04 vm05 ceph-mon[50927]: purged_snaps scrub starts 2026-03-10T05:30:04.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:04 vm05 ceph-mon[50927]: purged_snaps scrub ok 2026-03-10T05:30:04.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:04 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:04.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:04 vm05 ceph-mon[50927]: osd.1 [v2:192.168.123.101:6810/1960293858,v1:192.168.123.101:6811/1960293858] boot 2026-03-10T05:30:04.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:04 vm05 ceph-mon[50927]: osdmap e14: 3 total, 2 up, 3 in 2026-03-10T05:30:04.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:04 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:04.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:04 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:04.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:04 vm01 ceph-mon[47941]: purged_snaps scrub starts 2026-03-10T05:30:04.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:04 vm01 ceph-mon[47941]: purged_snaps scrub ok 2026-03-10T05:30:04.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:04 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:04.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:04 vm01 ceph-mon[47941]: osd.1 [v2:192.168.123.101:6810/1960293858,v1:192.168.123.101:6811/1960293858] boot 2026-03-10T05:30:04.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:04 vm01 ceph-mon[47941]: osdmap e14: 3 total, 2 up, 3 in 2026-03-10T05:30:04.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:04 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:04.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:04 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:04.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:04 vm02 ceph-mon[50473]: purged_snaps scrub starts 2026-03-10T05:30:04.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:04 vm02 ceph-mon[50473]: purged_snaps scrub ok 2026-03-10T05:30:04.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:04 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:04.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:04 vm02 ceph-mon[50473]: osd.1 [v2:192.168.123.101:6810/1960293858,v1:192.168.123.101:6811/1960293858] boot 2026-03-10T05:30:04.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:04 vm02 ceph-mon[50473]: osdmap e14: 3 total, 2 up, 3 in 2026-03-10T05:30:04.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:04 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T05:30:04.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:04 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:05.515 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:05 vm02 ceph-mon[50473]: pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:30:05.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:05 vm05 ceph-mon[50927]: pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:30:05.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:05 vm01 ceph-mon[47941]: pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-10T05:30:06.393 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:06 vm02 ceph-mon[50473]: osdmap e15: 3 total, 2 up, 3 in 2026-03-10T05:30:06.394 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:06 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:06.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:06 vm05 ceph-mon[50927]: osdmap e15: 3 total, 2 up, 3 in 2026-03-10T05:30:06.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:06 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:06.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:06 vm01 ceph-mon[47941]: osdmap e15: 3 total, 2 up, 3 in 2026-03-10T05:30:06.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:06 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:07.422 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:07 vm02 ceph-mon[50473]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:07.422 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:07 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T05:30:07.422 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:07 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:07.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:07 vm05 ceph-mon[50927]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:07.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:07 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T05:30:07.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:07 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:07.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:07 vm01 ceph-mon[47941]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:07.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:07 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T05:30:07.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:07 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:08.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:08 vm02 ceph-mon[50473]: Deploying daemon osd.2 on vm02 2026-03-10T05:30:08.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:08 vm05 ceph-mon[50927]: Deploying daemon osd.2 on vm02 2026-03-10T05:30:08.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:08 vm01 ceph-mon[47941]: Deploying daemon osd.2 on vm02 2026-03-10T05:30:09.684 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:09 vm02 ceph-mon[50473]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:09.684 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:09 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:09.684 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:09 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:09.684 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:09 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:09.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:09 vm05 ceph-mon[50927]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:09.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:09 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:09.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:09 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:09.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:09 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:09.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:09 vm01 ceph-mon[47941]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:09.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:09 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:09.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:09 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:09.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:09 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:10.229 INFO:teuthology.orchestra.run.vm02.stdout:Created osd(s) 2 on host 'vm02' 2026-03-10T05:30:10.303 DEBUG:teuthology.orchestra.run.vm02:osd.2> sudo journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.2.service 2026-03-10T05:30:10.304 INFO:tasks.cephadm:Deploying osd.3 on vm02 with /dev/vdd... 2026-03-10T05:30:10.305 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- lvm zap /dev/vdd 2026-03-10T05:30:10.605 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.b/config 2026-03-10T05:30:11.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:11.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:11.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:10 vm01 ceph-mon[47941]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:11.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:11.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:11.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:11.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:10 vm02 ceph-mon[50473]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:11.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:11.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.148 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:30:10 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-2[53628]: 2026-03-10T05:30:10.870+0000 7f5dd1924740 -1 osd.2 0 log_to_monitors true 2026-03-10T05:30:11.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:11.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:11.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:10 vm05 ceph-mon[50927]: pgmap v35: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:11.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:11.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.989 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:11 vm02 ceph-mon[50473]: from='osd.2 [v2:192.168.123.102:6800/1562513062,v1:192.168.123.102:6801/1562513062]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T05:30:11.989 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:11 vm02 ceph-mon[50473]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T05:30:11.989 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:11 vm02 ceph-mon[50473]: Detected new or changed devices on vm02 2026-03-10T05:30:11.989 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:11 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.989 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:11 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.989 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:11 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:11.989 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:11 vm02 ceph-mon[50473]: Adjusting osd_memory_target on vm02 to 4353M 2026-03-10T05:30:11.989 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:11 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:11.989 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:11 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:11.990 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:11 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:11.990 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:11 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:12.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:11 vm01 ceph-mon[47941]: from='osd.2 [v2:192.168.123.102:6800/1562513062,v1:192.168.123.102:6801/1562513062]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T05:30:12.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:11 vm01 ceph-mon[47941]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T05:30:12.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:11 vm01 ceph-mon[47941]: Detected new or changed devices on vm02 2026-03-10T05:30:12.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:12.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:12.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:12.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:11 vm01 ceph-mon[47941]: Adjusting osd_memory_target on vm02 to 4353M 2026-03-10T05:30:12.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:12.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:12.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:12.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:12.093 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:30:12.112 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch daemon add osd vm02:/dev/vdd 2026-03-10T05:30:12.293 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.b/config 2026-03-10T05:30:12.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:11 vm05 ceph-mon[50927]: from='osd.2 [v2:192.168.123.102:6800/1562513062,v1:192.168.123.102:6801/1562513062]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T05:30:12.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:11 vm05 ceph-mon[50927]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T05:30:12.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:11 vm05 ceph-mon[50927]: Detected new or changed devices on vm02 2026-03-10T05:30:12.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:11 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:12.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:11 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:12.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:11 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:12.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:11 vm05 ceph-mon[50927]: Adjusting osd_memory_target on vm02 to 4353M 2026-03-10T05:30:12.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:11 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:12.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:11 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:12.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:11 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:12.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:11 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:12 vm01 ceph-mon[47941]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T05:30:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:12 vm01 ceph-mon[47941]: osdmap e16: 3 total, 2 up, 3 in 2026-03-10T05:30:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:12 vm01 ceph-mon[47941]: from='osd.2 [v2:192.168.123.102:6800/1562513062,v1:192.168.123.102:6801/1562513062]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:12 vm01 ceph-mon[47941]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:12 vm01 ceph-mon[47941]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:13.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:12 vm02 ceph-mon[50473]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T05:30:13.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:12 vm02 ceph-mon[50473]: osdmap e16: 3 total, 2 up, 3 in 2026-03-10T05:30:13.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:13.149 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:12 vm02 ceph-mon[50473]: from='osd.2 [v2:192.168.123.102:6800/1562513062,v1:192.168.123.102:6801/1562513062]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:13.149 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:12 vm02 ceph-mon[50473]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:13.149 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:12 vm02 ceph-mon[50473]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:13.149 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:13.149 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:13.149 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:13.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:12 vm05 ceph-mon[50927]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T05:30:13.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:12 vm05 ceph-mon[50927]: osdmap e16: 3 total, 2 up, 3 in 2026-03-10T05:30:13.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:13.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:12 vm05 ceph-mon[50927]: from='osd.2 [v2:192.168.123.102:6800/1562513062,v1:192.168.123.102:6801/1562513062]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:13.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:12 vm05 ceph-mon[50927]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:13.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:12 vm05 ceph-mon[50927]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:13.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:13.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:13.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:14.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:13 vm01 ceph-mon[47941]: from='client.24176 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:14.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:13 vm01 ceph-mon[47941]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-10T05:30:14.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:13 vm01 ceph-mon[47941]: osdmap e17: 3 total, 2 up, 3 in 2026-03-10T05:30:14.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:13 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:14.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:13 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:14.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:13 vm01 ceph-mon[47941]: from='client.? 192.168.123.102:0/3856371827' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9916ed76-ed97-49db-960f-84cec5746a36"}]: dispatch 2026-03-10T05:30:14.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:13 vm01 ceph-mon[47941]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9916ed76-ed97-49db-960f-84cec5746a36"}]: dispatch 2026-03-10T05:30:14.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:13 vm01 ceph-mon[47941]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "9916ed76-ed97-49db-960f-84cec5746a36"}]': finished 2026-03-10T05:30:14.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:13 vm01 ceph-mon[47941]: osdmap e18: 4 total, 2 up, 4 in 2026-03-10T05:30:14.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:13 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:14.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:13 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:14.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:13 vm01 ceph-mon[47941]: from='osd.2 ' entity='osd.2' 2026-03-10T05:30:14.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:13 vm02 ceph-mon[50473]: from='client.24176 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:14.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:13 vm02 ceph-mon[50473]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-10T05:30:14.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:13 vm02 ceph-mon[50473]: osdmap e17: 3 total, 2 up, 3 in 2026-03-10T05:30:14.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:13 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:14.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:13 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:14.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:13 vm02 ceph-mon[50473]: from='client.? 192.168.123.102:0/3856371827' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9916ed76-ed97-49db-960f-84cec5746a36"}]: dispatch 2026-03-10T05:30:14.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:13 vm02 ceph-mon[50473]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9916ed76-ed97-49db-960f-84cec5746a36"}]: dispatch 2026-03-10T05:30:14.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:13 vm02 ceph-mon[50473]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "9916ed76-ed97-49db-960f-84cec5746a36"}]': finished 2026-03-10T05:30:14.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:13 vm02 ceph-mon[50473]: osdmap e18: 4 total, 2 up, 4 in 2026-03-10T05:30:14.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:13 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:14.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:13 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:14.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:13 vm02 ceph-mon[50473]: from='osd.2 ' entity='osd.2' 2026-03-10T05:30:14.148 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:30:13 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-2[53628]: 2026-03-10T05:30:13.740+0000 7f5dcd8a5640 -1 osd.2 0 waiting for initial osdmap 2026-03-10T05:30:14.148 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:30:13 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-2[53628]: 2026-03-10T05:30:13.753+0000 7f5dc8ece640 -1 osd.2 18 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T05:30:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:13 vm05 ceph-mon[50927]: from='client.24176 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:13 vm05 ceph-mon[50927]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-10T05:30:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:13 vm05 ceph-mon[50927]: osdmap e17: 3 total, 2 up, 3 in 2026-03-10T05:30:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:13 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:13 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:13 vm05 ceph-mon[50927]: from='client.? 192.168.123.102:0/3856371827' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9916ed76-ed97-49db-960f-84cec5746a36"}]: dispatch 2026-03-10T05:30:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:13 vm05 ceph-mon[50927]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9916ed76-ed97-49db-960f-84cec5746a36"}]: dispatch 2026-03-10T05:30:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:13 vm05 ceph-mon[50927]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "9916ed76-ed97-49db-960f-84cec5746a36"}]': finished 2026-03-10T05:30:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:13 vm05 ceph-mon[50927]: osdmap e18: 4 total, 2 up, 4 in 2026-03-10T05:30:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:13 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:13 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:13 vm05 ceph-mon[50927]: from='osd.2 ' entity='osd.2' 2026-03-10T05:30:15.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:14 vm05 ceph-mon[50927]: purged_snaps scrub starts 2026-03-10T05:30:15.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:14 vm05 ceph-mon[50927]: purged_snaps scrub ok 2026-03-10T05:30:15.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:14 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:15.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:14 vm05 ceph-mon[50927]: pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:15.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:14 vm05 ceph-mon[50927]: from='client.? 192.168.123.102:0/2009347855' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:15.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:14 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:15.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:14 vm01 ceph-mon[47941]: purged_snaps scrub starts 2026-03-10T05:30:15.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:14 vm01 ceph-mon[47941]: purged_snaps scrub ok 2026-03-10T05:30:15.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:15.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:14 vm01 ceph-mon[47941]: pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:15.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:14 vm01 ceph-mon[47941]: from='client.? 192.168.123.102:0/2009347855' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:15.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:15.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:14 vm02 ceph-mon[50473]: purged_snaps scrub starts 2026-03-10T05:30:15.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:14 vm02 ceph-mon[50473]: purged_snaps scrub ok 2026-03-10T05:30:15.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:14 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:15.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:14 vm02 ceph-mon[50473]: pgmap v40: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T05:30:15.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:14 vm02 ceph-mon[50473]: from='client.? 192.168.123.102:0/2009347855' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:15.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:14 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:16.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:15 vm05 ceph-mon[50927]: osd.2 [v2:192.168.123.102:6800/1562513062,v1:192.168.123.102:6801/1562513062] boot 2026-03-10T05:30:16.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:15 vm05 ceph-mon[50927]: osdmap e19: 4 total, 3 up, 4 in 2026-03-10T05:30:16.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:16.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:16.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T05:30:16.309 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:15 vm02 ceph-mon[50473]: osd.2 [v2:192.168.123.102:6800/1562513062,v1:192.168.123.102:6801/1562513062] boot 2026-03-10T05:30:16.309 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:15 vm02 ceph-mon[50473]: osdmap e19: 4 total, 3 up, 4 in 2026-03-10T05:30:16.309 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:16.309 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:16.310 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T05:30:16.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:15 vm01 ceph-mon[47941]: osd.2 [v2:192.168.123.102:6800/1562513062,v1:192.168.123.102:6801/1562513062] boot 2026-03-10T05:30:16.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:15 vm01 ceph-mon[47941]: osdmap e19: 4 total, 3 up, 4 in 2026-03-10T05:30:16.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T05:30:16.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:16.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T05:30:17.286 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:16 vm02 ceph-mon[50473]: pgmap v42: 0 pgs: ; 0 B data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:17.286 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:16 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T05:30:17.286 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:16 vm02 ceph-mon[50473]: osdmap e20: 4 total, 3 up, 4 in 2026-03-10T05:30:17.286 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:16 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:17.286 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:16 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T05:30:17.286 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 sudo[58247]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-10T05:30:17.286 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 sudo[58247]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T05:30:17.286 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 sudo[58247]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T05:30:17.286 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 sudo[58247]: pam_unix(sudo:session): session closed for user root 2026-03-10T05:30:17.286 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:30:17 vm02 sudo[58235]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-10T05:30:17.286 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:30:17 vm02 sudo[58235]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T05:30:17.286 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:30:17 vm02 sudo[58235]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T05:30:17.286 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:30:17 vm02 sudo[58235]: pam_unix(sudo:session): session closed for user root 2026-03-10T05:30:17.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:16 vm05 ceph-mon[50927]: pgmap v42: 0 pgs: ; 0 B data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:17.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:16 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T05:30:17.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:16 vm05 ceph-mon[50927]: osdmap e20: 4 total, 3 up, 4 in 2026-03-10T05:30:17.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:16 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:17.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:16 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T05:30:17.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 sudo[52275]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-10T05:30:17.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 sudo[52275]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T05:30:17.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 sudo[52275]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T05:30:17.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 sudo[52275]: pam_unix(sudo:session): session closed for user root 2026-03-10T05:30:17.338 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:30:17 vm01 sudo[65283]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-10T05:30:17.339 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:30:17 vm01 sudo[65283]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T05:30:17.339 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:30:17 vm01 sudo[65283]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T05:30:17.339 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:30:17 vm01 sudo[65283]: pam_unix(sudo:session): session closed for user root 2026-03-10T05:30:17.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:16 vm01 ceph-mon[47941]: pgmap v42: 0 pgs: ; 0 B data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:17.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:16 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T05:30:17.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:16 vm01 ceph-mon[47941]: osdmap e20: 4 total, 3 up, 4 in 2026-03-10T05:30:17.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:16 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:17.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:16 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T05:30:17.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 sudo[65291]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-10T05:30:17.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 sudo[65291]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T05:30:17.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 sudo[65291]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T05:30:17.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 sudo[65291]: pam_unix(sudo:session): session closed for user root 2026-03-10T05:30:17.339 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:30:17 vm01 sudo[65287]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vdd 2026-03-10T05:30:17.339 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:30:17 vm01 sudo[65287]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T05:30:17.339 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:30:17 vm01 sudo[65287]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T05:30:17.339 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:30:17 vm01 sudo[65287]: pam_unix(sudo:session): session closed for user root 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: osdmap e21: 4 total, 3 up, 4 in 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:18.096 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:17 vm02 ceph-mon[50473]: Deploying daemon osd.3 on vm02 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: osdmap e21: 4 total, 3 up, 4 in 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:30:18.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:30:18.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:30:18.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T05:30:18.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T05:30:18.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:18.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:17 vm05 ceph-mon[50927]: Deploying daemon osd.3 on vm02 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: osdmap e21: 4 total, 3 up, 4 in 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:18.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:17 vm01 ceph-mon[47941]: Deploying daemon osd.3 on vm02 2026-03-10T05:30:19.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:18 vm02 ceph-mon[50473]: pgmap v45: 1 pgs: 1 unknown; 0 B data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:19.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:18 vm02 ceph-mon[50473]: mgrmap e13: x(active, since 68s) 2026-03-10T05:30:19.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:18 vm02 ceph-mon[50473]: osdmap e22: 4 total, 3 up, 4 in 2026-03-10T05:30:19.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:18 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:19.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:18 vm05 ceph-mon[50927]: pgmap v45: 1 pgs: 1 unknown; 0 B data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:19.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:18 vm05 ceph-mon[50927]: mgrmap e13: x(active, since 68s) 2026-03-10T05:30:19.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:18 vm05 ceph-mon[50927]: osdmap e22: 4 total, 3 up, 4 in 2026-03-10T05:30:19.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:18 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:19.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:18 vm01 ceph-mon[47941]: pgmap v45: 1 pgs: 1 unknown; 0 B data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:19.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:18 vm01 ceph-mon[47941]: mgrmap e13: x(active, since 68s) 2026-03-10T05:30:19.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:18 vm01 ceph-mon[47941]: osdmap e22: 4 total, 3 up, 4 in 2026-03-10T05:30:19.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:18 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:20.149 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:19 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:20.149 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:19 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:20.149 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:19 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:20.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:19 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:20.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:19 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:20.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:19 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:20.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:19 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:20.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:19 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:20.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:19 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:20.665 INFO:teuthology.orchestra.run.vm02.stdout:Created osd(s) 3 on host 'vm02' 2026-03-10T05:30:20.729 DEBUG:teuthology.orchestra.run.vm02:osd.3> sudo journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.3.service 2026-03-10T05:30:20.730 INFO:tasks.cephadm:Deploying osd.4 on vm02 with /dev/vdc... 2026-03-10T05:30:20.730 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- lvm zap /dev/vdc 2026-03-10T05:30:21.020 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.b/config 2026-03-10T05:30:21.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:21 vm02 ceph-mon[50473]: pgmap v47: 1 pgs: 1 unknown; 0 B data, 480 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:21.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:21 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:21 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:21 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:21.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:21 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:21.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:21 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:21 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:21.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:21 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:21 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:21 vm05 ceph-mon[50927]: pgmap v47: 1 pgs: 1 unknown; 0 B data, 480 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:21.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:21 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:21 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:21 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:21.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:21 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:21.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:21 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:21 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:21.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:21 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:21 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:21 vm01 ceph-mon[47941]: pgmap v47: 1 pgs: 1 unknown; 0 B data, 480 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:21.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:21 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:21 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:21 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:21.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:21 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:21.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:21 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:21 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:21.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:21 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:21 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:21.902 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:30:21 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-3[58807]: 2026-03-10T05:30:21.657+0000 7fa51720c740 -1 osd.3 0 log_to_monitors true 2026-03-10T05:30:22.723 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:30:22.743 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch daemon add osd vm02:/dev/vdc 2026-03-10T05:30:22.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:22 vm02 ceph-mon[50473]: Detected new or changed devices on vm02 2026-03-10T05:30:22.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:22 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:22.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:22 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:22.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:22 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:22.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:22 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:22.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:22 vm02 ceph-mon[50473]: Adjusting osd_memory_target on vm02 to 2176M 2026-03-10T05:30:22.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:22 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:22.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:22 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:22.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:22 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:22.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:22 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:22.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:22 vm02 ceph-mon[50473]: from='osd.3 [v2:192.168.123.102:6808/914771124,v1:192.168.123.102:6809/914771124]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T05:30:22.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:22 vm02 ceph-mon[50473]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T05:30:22.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:22 vm02 ceph-mon[50473]: pgmap v48: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:22.917 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.b/config 2026-03-10T05:30:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:22 vm05 ceph-mon[50927]: Detected new or changed devices on vm02 2026-03-10T05:30:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:22 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:22 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:22 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:22 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:22 vm05 ceph-mon[50927]: Adjusting osd_memory_target on vm02 to 2176M 2026-03-10T05:30:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:22 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:22 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:22 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:22 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:22 vm05 ceph-mon[50927]: from='osd.3 [v2:192.168.123.102:6808/914771124,v1:192.168.123.102:6809/914771124]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T05:30:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:22 vm05 ceph-mon[50927]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T05:30:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:22 vm05 ceph-mon[50927]: pgmap v48: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:22 vm01 ceph-mon[47941]: Detected new or changed devices on vm02 2026-03-10T05:30:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:22 vm01 ceph-mon[47941]: Adjusting osd_memory_target on vm02 to 2176M 2026-03-10T05:30:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:22 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:23.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:22 vm01 ceph-mon[47941]: from='osd.3 [v2:192.168.123.102:6808/914771124,v1:192.168.123.102:6809/914771124]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T05:30:23.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:22 vm01 ceph-mon[47941]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T05:30:23.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:22 vm01 ceph-mon[47941]: pgmap v48: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:23.993 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:30:23 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-3[58807]: 2026-03-10T05:30:23.730+0000 7fa51318d640 -1 osd.3 0 waiting for initial osdmap 2026-03-10T05:30:23.993 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:30:23 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-3[58807]: 2026-03-10T05:30:23.738+0000 7fa50efb7640 -1 osd.3 24 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T05:30:23.993 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:23 vm02 ceph-mon[50473]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T05:30:23.993 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:23 vm02 ceph-mon[50473]: osdmap e23: 4 total, 3 up, 4 in 2026-03-10T05:30:23.993 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:23 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:23.993 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:23 vm02 ceph-mon[50473]: from='osd.3 [v2:192.168.123.102:6808/914771124,v1:192.168.123.102:6809/914771124]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:23.993 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:23 vm02 ceph-mon[50473]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:23.993 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:23 vm02 ceph-mon[50473]: from='client.24212 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:23.993 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:23 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:23.993 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:23 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:23.993 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:23 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:24.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:23 vm05 ceph-mon[50927]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T05:30:24.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:23 vm05 ceph-mon[50927]: osdmap e23: 4 total, 3 up, 4 in 2026-03-10T05:30:24.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:23 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:24.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:23 vm05 ceph-mon[50927]: from='osd.3 [v2:192.168.123.102:6808/914771124,v1:192.168.123.102:6809/914771124]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:24.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:23 vm05 ceph-mon[50927]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:24.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:23 vm05 ceph-mon[50927]: from='client.24212 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:24.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:23 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:24.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:23 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:24.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:23 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:24.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:23 vm01 ceph-mon[47941]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T05:30:24.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:23 vm01 ceph-mon[47941]: osdmap e23: 4 total, 3 up, 4 in 2026-03-10T05:30:24.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:23 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:24.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:23 vm01 ceph-mon[47941]: from='osd.3 [v2:192.168.123.102:6808/914771124,v1:192.168.123.102:6809/914771124]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:24.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:23 vm01 ceph-mon[47941]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:24.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:23 vm01 ceph-mon[47941]: from='client.24212 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:24.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:23 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:24.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:23 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:24.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:23 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: purged_snaps scrub starts 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: purged_snaps scrub ok 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: osdmap e24: 4 total, 3 up, 4 in 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.102:0/2675393652' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "cb979506-99a8-4950-b248-04eeab920b11"}]: dispatch 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "cb979506-99a8-4950-b248-04eeab920b11"}]: dispatch 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "cb979506-99a8-4950-b248-04eeab920b11"}]': finished 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: osd.3 [v2:192.168.123.102:6808/914771124,v1:192.168.123.102:6809/914771124] boot 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: osdmap e25: 5 total, 4 up, 5 in 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:25.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:24 vm05 ceph-mon[50927]: from='client.? 192.168.123.102:0/3955037543' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: purged_snaps scrub starts 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: purged_snaps scrub ok 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: osdmap e24: 4 total, 3 up, 4 in 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: from='client.? 192.168.123.102:0/2675393652' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "cb979506-99a8-4950-b248-04eeab920b11"}]: dispatch 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "cb979506-99a8-4950-b248-04eeab920b11"}]: dispatch 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "cb979506-99a8-4950-b248-04eeab920b11"}]': finished 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: osd.3 [v2:192.168.123.102:6808/914771124,v1:192.168.123.102:6809/914771124] boot 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: osdmap e25: 5 total, 4 up, 5 in 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:25.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:24 vm01 ceph-mon[47941]: from='client.? 192.168.123.102:0/3955037543' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:25.147 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: purged_snaps scrub starts 2026-03-10T05:30:25.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: purged_snaps scrub ok 2026-03-10T05:30:25.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-10T05:30:25.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: osdmap e24: 4 total, 3 up, 4 in 2026-03-10T05:30:25.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:25.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:25.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T05:30:25.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: from='client.? 192.168.123.102:0/2675393652' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "cb979506-99a8-4950-b248-04eeab920b11"}]: dispatch 2026-03-10T05:30:25.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "cb979506-99a8-4950-b248-04eeab920b11"}]: dispatch 2026-03-10T05:30:25.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "cb979506-99a8-4950-b248-04eeab920b11"}]': finished 2026-03-10T05:30:25.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: osd.3 [v2:192.168.123.102:6808/914771124,v1:192.168.123.102:6809/914771124] boot 2026-03-10T05:30:25.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: osdmap e25: 5 total, 4 up, 5 in 2026-03-10T05:30:25.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T05:30:25.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:25.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:24 vm02 ceph-mon[50473]: from='client.? 192.168.123.102:0/3955037543' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:26.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:25 vm05 ceph-mon[50927]: osdmap e26: 5 total, 4 up, 5 in 2026-03-10T05:30:26.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:25 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:26.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:25 vm01 ceph-mon[47941]: osdmap e26: 5 total, 4 up, 5 in 2026-03-10T05:30:26.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:25 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:26.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:25 vm02 ceph-mon[50473]: osdmap e26: 5 total, 4 up, 5 in 2026-03-10T05:30:26.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:25 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:27.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:26 vm05 ceph-mon[50927]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T05:30:27.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:26 vm05 ceph-mon[50927]: osdmap e27: 5 total, 4 up, 5 in 2026-03-10T05:30:27.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:26 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:27.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:26 vm01 ceph-mon[47941]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T05:30:27.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:26 vm01 ceph-mon[47941]: osdmap e27: 5 total, 4 up, 5 in 2026-03-10T05:30:27.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:26 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:27.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:26 vm02 ceph-mon[50473]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T05:30:27.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:26 vm02 ceph-mon[50473]: osdmap e27: 5 total, 4 up, 5 in 2026-03-10T05:30:27.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:26 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:29.110 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:28 vm02 ceph-mon[50473]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T05:30:29.110 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:28 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T05:30:29.110 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:28 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:29.110 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:28 vm02 ceph-mon[50473]: Deploying daemon osd.4 on vm02 2026-03-10T05:30:29.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:28 vm05 ceph-mon[50927]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T05:30:29.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:28 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T05:30:29.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:28 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:29.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:28 vm05 ceph-mon[50927]: Deploying daemon osd.4 on vm02 2026-03-10T05:30:29.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:28 vm01 ceph-mon[47941]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T05:30:29.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:28 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T05:30:29.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:28 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:29.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:28 vm01 ceph-mon[47941]: Deploying daemon osd.4 on vm02 2026-03-10T05:30:31.256 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:30 vm02 ceph-mon[50473]: pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T05:30:31.256 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:30 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:31.256 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:30 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.256 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:30 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.256 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:30 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.256 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:30 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.256 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:30 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:31.256 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:30 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:31.256 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:30 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.295 INFO:teuthology.orchestra.run.vm02.stdout:Created osd(s) 4 on host 'vm02' 2026-03-10T05:30:31.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:30 vm05 ceph-mon[50927]: pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T05:30:31.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:30 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:31.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:30 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:30 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:30 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:30 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:30 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:31.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:30 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:31.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:30 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:30 vm01 ceph-mon[47941]: pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T05:30:31.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:30 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:31.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:30 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:30 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:30 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:30 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:30 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:31.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:30 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:31.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:30 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:31.352 DEBUG:teuthology.orchestra.run.vm02:osd.4> sudo journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.4.service 2026-03-10T05:30:31.353 INFO:tasks.cephadm:Deploying osd.5 on vm05 with /dev/vde... 2026-03-10T05:30:31.353 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- lvm zap /dev/vde 2026-03-10T05:30:31.534 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.c/config 2026-03-10T05:30:31.544 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:30:31 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-4[63840]: 2026-03-10T05:30:31.493+0000 7fb2bcc36740 -1 Falling back to public interface 2026-03-10T05:30:32.251 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:31 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:32.252 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:31 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:32.252 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:31 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:32.256 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:31 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:32.256 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:31 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:32.256 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:31 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:32.328 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:30:32.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:31 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:32.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:31 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:32.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:31 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:32.344 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch daemon add osd vm05:/dev/vde 2026-03-10T05:30:32.520 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.c/config 2026-03-10T05:30:32.898 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:30:32 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-4[63840]: 2026-03-10T05:30:32.587+0000 7fb2bcc36740 -1 osd.4 0 log_to_monitors true 2026-03-10T05:30:33.234 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 57 KiB/s, 0 objects/s recovering 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 57 KiB/s, 0 objects/s recovering 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: Detected new or changed devices on vm02 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: Adjusting osd_memory_target on vm02 to 1451M 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: from='osd.4 [v2:192.168.123.102:6816/4203989344,v1:192.168.123.102:6817/4203989344]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:33.510 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:33 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: Detected new or changed devices on vm02 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: Adjusting osd_memory_target on vm02 to 1451M 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: from='osd.4 [v2:192.168.123.102:6816/4203989344,v1:192.168.123.102:6817/4203989344]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:33.546 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:33 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:33.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 57 KiB/s, 0 objects/s recovering 2026-03-10T05:30:33.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: Detected new or changed devices on vm02 2026-03-10T05:30:33.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:33.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:33.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:33.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:33.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:33.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: Adjusting osd_memory_target on vm02 to 1451M 2026-03-10T05:30:33.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:33.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:33.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:33.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:33.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: from='osd.4 [v2:192.168.123.102:6816/4203989344,v1:192.168.123.102:6817/4203989344]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T05:30:33.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T05:30:33.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:33.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:33.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:33 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:33.897 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:30:33 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-4[63840]: 2026-03-10T05:30:33.506+0000 7fb2b8bb7640 -1 osd.4 0 waiting for initial osdmap 2026-03-10T05:30:33.897 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:30:33 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-4[63840]: 2026-03-10T05:30:33.512+0000 7fb2b49e1640 -1 osd.4 29 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T05:30:34.301 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: from='client.14334 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:34.301 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T05:30:34.301 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: osdmap e28: 5 total, 4 up, 5 in 2026-03-10T05:30:34.301 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:34.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: from='osd.4 [v2:192.168.123.102:6816/4203989344,v1:192.168.123.102:6817/4203989344]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:34.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:34.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/330015760' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4615618e-f690-4722-b0c7-d3d66c66ee61"}]: dispatch 2026-03-10T05:30:34.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4615618e-f690-4722-b0c7-d3d66c66ee61"}]: dispatch 2026-03-10T05:30:34.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-10T05:30:34.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4615618e-f690-4722-b0c7-d3d66c66ee61"}]': finished 2026-03-10T05:30:34.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: osdmap e29: 6 total, 4 up, 6 in 2026-03-10T05:30:34.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:34.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:34.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:34.302 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:34 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1808825486' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:34.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: from='client.14334 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:34.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T05:30:34.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: osdmap e28: 5 total, 4 up, 5 in 2026-03-10T05:30:34.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:34.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: from='osd.4 [v2:192.168.123.102:6816/4203989344,v1:192.168.123.102:6817/4203989344]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:34.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:34.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/330015760' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4615618e-f690-4722-b0c7-d3d66c66ee61"}]: dispatch 2026-03-10T05:30:34.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4615618e-f690-4722-b0c7-d3d66c66ee61"}]: dispatch 2026-03-10T05:30:34.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-10T05:30:34.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4615618e-f690-4722-b0c7-d3d66c66ee61"}]': finished 2026-03-10T05:30:34.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: osdmap e29: 6 total, 4 up, 6 in 2026-03-10T05:30:34.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:34.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:34.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:34.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:34 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1808825486' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:34.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: from='client.14334 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:34.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T05:30:34.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: osdmap e28: 5 total, 4 up, 5 in 2026-03-10T05:30:34.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:34.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: from='osd.4 [v2:192.168.123.102:6816/4203989344,v1:192.168.123.102:6817/4203989344]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:34.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-10T05:30:34.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/330015760' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4615618e-f690-4722-b0c7-d3d66c66ee61"}]: dispatch 2026-03-10T05:30:34.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4615618e-f690-4722-b0c7-d3d66c66ee61"}]: dispatch 2026-03-10T05:30:34.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-10T05:30:34.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4615618e-f690-4722-b0c7-d3d66c66ee61"}]': finished 2026-03-10T05:30:34.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: osdmap e29: 6 total, 4 up, 6 in 2026-03-10T05:30:34.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:34.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:34.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:34.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:34 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1808825486' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:35.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:35 vm05 ceph-mon[50927]: pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 57 KiB/s, 0 objects/s recovering 2026-03-10T05:30:35.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:35 vm05 ceph-mon[50927]: osd.4 [v2:192.168.123.102:6816/4203989344,v1:192.168.123.102:6817/4203989344] boot 2026-03-10T05:30:35.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:35 vm05 ceph-mon[50927]: osdmap e30: 6 total, 5 up, 6 in 2026-03-10T05:30:35.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:35.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:35 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:35.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:35 vm01 ceph-mon[47941]: pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 57 KiB/s, 0 objects/s recovering 2026-03-10T05:30:35.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:35 vm01 ceph-mon[47941]: osd.4 [v2:192.168.123.102:6816/4203989344,v1:192.168.123.102:6817/4203989344] boot 2026-03-10T05:30:35.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:35 vm01 ceph-mon[47941]: osdmap e30: 6 total, 5 up, 6 in 2026-03-10T05:30:35.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:35.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:35 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:35.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:35 vm02 ceph-mon[50473]: pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 57 KiB/s, 0 objects/s recovering 2026-03-10T05:30:35.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:35 vm02 ceph-mon[50473]: osd.4 [v2:192.168.123.102:6816/4203989344,v1:192.168.123.102:6817/4203989344] boot 2026-03-10T05:30:35.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:35 vm02 ceph-mon[50473]: osdmap e30: 6 total, 5 up, 6 in 2026-03-10T05:30:35.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T05:30:35.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:35 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:36.676 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:36 vm05 ceph-mon[50927]: purged_snaps scrub starts 2026-03-10T05:30:36.676 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:36 vm05 ceph-mon[50927]: purged_snaps scrub ok 2026-03-10T05:30:36.676 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:36 vm05 ceph-mon[50927]: osdmap e31: 6 total, 5 up, 6 in 2026-03-10T05:30:36.676 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:36 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:36.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:36 vm01 ceph-mon[47941]: purged_snaps scrub starts 2026-03-10T05:30:36.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:36 vm01 ceph-mon[47941]: purged_snaps scrub ok 2026-03-10T05:30:36.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:36 vm01 ceph-mon[47941]: osdmap e31: 6 total, 5 up, 6 in 2026-03-10T05:30:36.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:36 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:36.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:36 vm02 ceph-mon[50473]: purged_snaps scrub starts 2026-03-10T05:30:36.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:36 vm02 ceph-mon[50473]: purged_snaps scrub ok 2026-03-10T05:30:36.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:36 vm02 ceph-mon[50473]: osdmap e31: 6 total, 5 up, 6 in 2026-03-10T05:30:36.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:36 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:37.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:37 vm02 ceph-mon[50473]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:38.000 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:37 vm05 ceph-mon[50927]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:38.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:37 vm01 ceph-mon[47941]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:38.774 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:38 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T05:30:38.774 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:38 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:38.774 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:38 vm05 ceph-mon[50927]: Deploying daemon osd.5 on vm05 2026-03-10T05:30:38.774 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:38 vm05 ceph-mon[50927]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:38.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:38 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T05:30:38.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:38 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:38.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:38 vm02 ceph-mon[50473]: Deploying daemon osd.5 on vm05 2026-03-10T05:30:38.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:38 vm02 ceph-mon[50473]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:39.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:38 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T05:30:39.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:38 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:39.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:38 vm01 ceph-mon[47941]: Deploying daemon osd.5 on vm05 2026-03-10T05:30:39.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:38 vm01 ceph-mon[47941]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:40.924 INFO:teuthology.orchestra.run.vm05.stdout:Created osd(s) 5 on host 'vm05' 2026-03-10T05:30:40.984 DEBUG:teuthology.orchestra.run.vm05:osd.5> sudo journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.5.service 2026-03-10T05:30:40.986 INFO:tasks.cephadm:Deploying osd.6 on vm05 with /dev/vdd... 2026-03-10T05:30:40.986 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- lvm zap /dev/vdd 2026-03-10T05:30:41.202 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:40 vm05 ceph-mon[50927]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:41.202 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:40 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:41.202 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:40 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:41.202 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:40 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:41.202 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:40 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:41.202 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:40 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:41.202 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:40 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:41.202 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:40 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:41.202 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:40 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:41.292 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.c/config 2026-03-10T05:30:41.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:40 vm01 ceph-mon[47941]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:41.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:40 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:41.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:40 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:41.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:40 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:41.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:40 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:41.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:40 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:41.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:40 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:41.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:40 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:41.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:40 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:41.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:40 vm02 ceph-mon[50473]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:41.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:40 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:41.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:40 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:41.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:40 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:41.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:40 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:41.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:40 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:41.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:40 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:41.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:40 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:41.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:40 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:42.078 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:41 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:42.078 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:41 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:42.078 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:41 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:42.078 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:41 vm05 ceph-mon[50927]: from='osd.5 [v2:192.168.123.105:6800/576190648,v1:192.168.123.105:6801/576190648]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T05:30:42.078 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:41 vm05 ceph-mon[50927]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T05:30:42.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:41 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:42.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:41 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:42.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:41 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:42.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:41 vm01 ceph-mon[47941]: from='osd.5 [v2:192.168.123.105:6800/576190648,v1:192.168.123.105:6801/576190648]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T05:30:42.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:41 vm01 ceph-mon[47941]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T05:30:42.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:41 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:42.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:41 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:42.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:41 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:42.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:41 vm02 ceph-mon[50473]: from='osd.5 [v2:192.168.123.105:6800/576190648,v1:192.168.123.105:6801/576190648]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T05:30:42.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:41 vm02 ceph-mon[50473]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T05:30:42.636 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:30:42.654 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch daemon add osd vm05:/dev/vdd 2026-03-10T05:30:42.819 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.c/config 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: osdmap e32: 6 total, 5 up, 6 in 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: Detected new or changed devices on vm05 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: from='osd.5 [v2:192.168.123.105:6800/576190648,v1:192.168.123.105:6801/576190648]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: Adjusting osd_memory_target on vm05 to 4353M 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:42 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:43.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:43.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T05:30:43.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: osdmap e32: 6 total, 5 up, 6 in 2026-03-10T05:30:43.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:43.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: Detected new or changed devices on vm05 2026-03-10T05:30:43.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: from='osd.5 [v2:192.168.123.105:6800/576190648,v1:192.168.123.105:6801/576190648]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:30:43.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:30:43.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:43.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:43.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:43.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: Adjusting osd_memory_target on vm05 to 4353M 2026-03-10T05:30:43.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:43.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:43.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:43.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:42 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: osdmap e32: 6 total, 5 up, 6 in 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: Detected new or changed devices on vm05 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: from='osd.5 [v2:192.168.123.105:6800/576190648,v1:192.168.123.105:6801/576190648]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: Adjusting osd_memory_target on vm05 to 4353M 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:42 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:44.001 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:43 vm05 ceph-mon[50927]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-10T05:30:44.001 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:43 vm05 ceph-mon[50927]: osdmap e33: 6 total, 5 up, 6 in 2026-03-10T05:30:44.001 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:43 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:44.001 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:43 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:44.001 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:43 vm05 ceph-mon[50927]: from='client.24238 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:44.001 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:43 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:44.001 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:43 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:44.001 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:43 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:44.001 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:43 vm05 ceph-mon[50927]: from='osd.5 ' entity='osd.5' 2026-03-10T05:30:44.001 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:43 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1337999228' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "33f0f20d-1d48-4480-9474-610fa90103f6"}]: dispatch 2026-03-10T05:30:44.001 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:43 vm05 ceph-mon[50927]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "33f0f20d-1d48-4480-9474-610fa90103f6"}]: dispatch 2026-03-10T05:30:44.001 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:43 vm05 ceph-mon[50927]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "33f0f20d-1d48-4480-9474-610fa90103f6"}]': finished 2026-03-10T05:30:44.001 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 10 05:30:43 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-5[54445]: 2026-03-10T05:30:43.851+0000 7fc7d276e640 -1 osd.5 0 waiting for initial osdmap 2026-03-10T05:30:44.001 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 10 05:30:43 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-5[54445]: 2026-03-10T05:30:43.862+0000 7fc7cd584640 -1 osd.5 33 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T05:30:44.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:43 vm01 ceph-mon[47941]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-10T05:30:44.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:43 vm01 ceph-mon[47941]: osdmap e33: 6 total, 5 up, 6 in 2026-03-10T05:30:44.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:43 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:44.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:43 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:44.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:43 vm01 ceph-mon[47941]: from='client.24238 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:44.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:43 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:44.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:43 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:44.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:43 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:44.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:43 vm01 ceph-mon[47941]: from='osd.5 ' entity='osd.5' 2026-03-10T05:30:44.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:43 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1337999228' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "33f0f20d-1d48-4480-9474-610fa90103f6"}]: dispatch 2026-03-10T05:30:44.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:43 vm01 ceph-mon[47941]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "33f0f20d-1d48-4480-9474-610fa90103f6"}]: dispatch 2026-03-10T05:30:44.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:43 vm01 ceph-mon[47941]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "33f0f20d-1d48-4480-9474-610fa90103f6"}]': finished 2026-03-10T05:30:44.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:43 vm02 ceph-mon[50473]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-10T05:30:44.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:43 vm02 ceph-mon[50473]: osdmap e33: 6 total, 5 up, 6 in 2026-03-10T05:30:44.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:43 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:44.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:43 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:44.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:43 vm02 ceph-mon[50473]: from='client.24238 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:44.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:43 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:44.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:43 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:44.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:43 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:44.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:43 vm02 ceph-mon[50473]: from='osd.5 ' entity='osd.5' 2026-03-10T05:30:44.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:43 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1337999228' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "33f0f20d-1d48-4480-9474-610fa90103f6"}]: dispatch 2026-03-10T05:30:44.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:43 vm02 ceph-mon[50473]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "33f0f20d-1d48-4480-9474-610fa90103f6"}]: dispatch 2026-03-10T05:30:44.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:43 vm02 ceph-mon[50473]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "33f0f20d-1d48-4480-9474-610fa90103f6"}]': finished 2026-03-10T05:30:45.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:44 vm05 ceph-mon[50927]: purged_snaps scrub starts 2026-03-10T05:30:45.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:44 vm05 ceph-mon[50927]: purged_snaps scrub ok 2026-03-10T05:30:45.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:44 vm05 ceph-mon[50927]: pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:45.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:44 vm05 ceph-mon[50927]: osd.5 [v2:192.168.123.105:6800/576190648,v1:192.168.123.105:6801/576190648] boot 2026-03-10T05:30:45.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:44 vm05 ceph-mon[50927]: osdmap e34: 7 total, 6 up, 7 in 2026-03-10T05:30:45.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:44 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:45.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:44 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:45.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:44 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/4036316046' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:45.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:44 vm05 ceph-mon[50927]: osdmap e35: 7 total, 6 up, 7 in 2026-03-10T05:30:45.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:44 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:45.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:44 vm01 ceph-mon[47941]: purged_snaps scrub starts 2026-03-10T05:30:45.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:44 vm01 ceph-mon[47941]: purged_snaps scrub ok 2026-03-10T05:30:45.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:44 vm01 ceph-mon[47941]: pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:45.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:44 vm01 ceph-mon[47941]: osd.5 [v2:192.168.123.105:6800/576190648,v1:192.168.123.105:6801/576190648] boot 2026-03-10T05:30:45.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:44 vm01 ceph-mon[47941]: osdmap e34: 7 total, 6 up, 7 in 2026-03-10T05:30:45.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:44 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:45.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:44 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:45.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:44 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/4036316046' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:45.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:44 vm01 ceph-mon[47941]: osdmap e35: 7 total, 6 up, 7 in 2026-03-10T05:30:45.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:44 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:45.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:44 vm02 ceph-mon[50473]: purged_snaps scrub starts 2026-03-10T05:30:45.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:44 vm02 ceph-mon[50473]: purged_snaps scrub ok 2026-03-10T05:30:45.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:44 vm02 ceph-mon[50473]: pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T05:30:45.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:44 vm02 ceph-mon[50473]: osd.5 [v2:192.168.123.105:6800/576190648,v1:192.168.123.105:6801/576190648] boot 2026-03-10T05:30:45.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:44 vm02 ceph-mon[50473]: osdmap e34: 7 total, 6 up, 7 in 2026-03-10T05:30:45.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:44 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T05:30:45.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:44 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:45.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:44 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/4036316046' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:45.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:44 vm02 ceph-mon[50473]: osdmap e35: 7 total, 6 up, 7 in 2026-03-10T05:30:45.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:44 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:47.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:46 vm05 ceph-mon[50927]: pgmap v73: 1 pgs: 1 remapped+peering; 449 KiB data, 200 MiB used, 120 GiB / 120 GiB avail 2026-03-10T05:30:47.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:46 vm05 ceph-mon[50927]: osdmap e36: 7 total, 6 up, 7 in 2026-03-10T05:30:47.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:46 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:47.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:46 vm01 ceph-mon[47941]: pgmap v73: 1 pgs: 1 remapped+peering; 449 KiB data, 200 MiB used, 120 GiB / 120 GiB avail 2026-03-10T05:30:47.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:46 vm01 ceph-mon[47941]: osdmap e36: 7 total, 6 up, 7 in 2026-03-10T05:30:47.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:46 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:47.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:46 vm02 ceph-mon[50473]: pgmap v73: 1 pgs: 1 remapped+peering; 449 KiB data, 200 MiB used, 120 GiB / 120 GiB avail 2026-03-10T05:30:47.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:46 vm02 ceph-mon[50473]: osdmap e36: 7 total, 6 up, 7 in 2026-03-10T05:30:47.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:46 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:49.185 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:48 vm05 ceph-mon[50927]: pgmap v75: 1 pgs: 1 remapped+peering; 449 KiB data, 200 MiB used, 120 GiB / 120 GiB avail 2026-03-10T05:30:49.185 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:48 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-10T05:30:49.185 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:48 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:49.185 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:48 vm05 ceph-mon[50927]: Deploying daemon osd.6 on vm05 2026-03-10T05:30:49.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:48 vm01 ceph-mon[47941]: pgmap v75: 1 pgs: 1 remapped+peering; 449 KiB data, 200 MiB used, 120 GiB / 120 GiB avail 2026-03-10T05:30:49.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:48 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-10T05:30:49.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:48 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:49.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:48 vm01 ceph-mon[47941]: Deploying daemon osd.6 on vm05 2026-03-10T05:30:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:48 vm02 ceph-mon[50473]: pgmap v75: 1 pgs: 1 remapped+peering; 449 KiB data, 200 MiB used, 120 GiB / 120 GiB avail 2026-03-10T05:30:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:48 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-10T05:30:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:48 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:48 vm02 ceph-mon[50473]: Deploying daemon osd.6 on vm05 2026-03-10T05:30:51.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:50 vm05 ceph-mon[50927]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T05:30:51.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:50 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:51.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:50 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:50 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:50 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:50 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:50 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:51.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:50 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:51.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:50 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:50 vm01 ceph-mon[47941]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T05:30:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:50 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:50 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:50 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:50 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:50 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:50 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:51.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:50 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:51.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:50 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.381 INFO:teuthology.orchestra.run.vm05.stdout:Created osd(s) 6 on host 'vm05' 2026-03-10T05:30:51.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:50 vm02 ceph-mon[50473]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T05:30:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:50 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:50 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:50 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:50 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:50 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:50 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:50 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:50 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:51.439 DEBUG:teuthology.orchestra.run.vm05:osd.6> sudo journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.6.service 2026-03-10T05:30:51.441 INFO:tasks.cephadm:Deploying osd.7 on vm05 with /dev/vdc... 2026-03-10T05:30:51.441 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- lvm zap /dev/vdc 2026-03-10T05:30:51.761 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.c/config 2026-03-10T05:30:52.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:51 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:52.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:51 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:52.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:51 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:52.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:51 vm05 ceph-mon[50927]: from='osd.6 [v2:192.168.123.105:6808/3164767821,v1:192.168.123.105:6809/3164767821]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T05:30:52.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:51 vm05 ceph-mon[50927]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T05:30:52.306 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:30:51 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-6[59586]: 2026-03-10T05:30:51.921+0000 7f3ecef2a740 -1 osd.6 0 log_to_monitors true 2026-03-10T05:30:52.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:51 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:52.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:51 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:52.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:51 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:52.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:51 vm01 ceph-mon[47941]: from='osd.6 [v2:192.168.123.105:6808/3164767821,v1:192.168.123.105:6809/3164767821]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T05:30:52.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:51 vm01 ceph-mon[47941]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T05:30:52.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:51 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:30:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:51 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:51 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:51 vm02 ceph-mon[50473]: from='osd.6 [v2:192.168.123.105:6808/3164767821,v1:192.168.123.105:6809/3164767821]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T05:30:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:51 vm02 ceph-mon[50473]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T05:30:53.158 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:30:53.177 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch daemon add osd vm05:/dev/vdc 2026-03-10T05:30:53.352 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.c/config 2026-03-10T05:30:53.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail; 57 KiB/s, 0 objects/s recovering 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: osdmap e37: 7 total, 6 up, 7 in 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: from='osd.6 [v2:192.168.123.105:6808/3164767821,v1:192.168.123.105:6809/3164767821]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: Detected new or changed devices on vm05 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: Adjusting osd_memory_target on vm05 to 2176M 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:53 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:53.790 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail; 57 KiB/s, 0 objects/s recovering 2026-03-10T05:30:53.790 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-10T05:30:53.790 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: osdmap e37: 7 total, 6 up, 7 in 2026-03-10T05:30:53.790 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:53.790 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:30:53.790 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: from='osd.6 [v2:192.168.123.105:6808/3164767821,v1:192.168.123.105:6809/3164767821]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:30:53.790 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: Detected new or changed devices on vm05 2026-03-10T05:30:53.790 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:53.791 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:53.791 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:53.791 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:53.791 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: Adjusting osd_memory_target on vm05 to 2176M 2026-03-10T05:30:53.791 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:53.791 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:53.791 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:53.791 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:53 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:53.791 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:30:53 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-6[59586]: 2026-03-10T05:30:53.396+0000 7f3ecb6be640 -1 osd.6 0 waiting for initial osdmap 2026-03-10T05:30:53.791 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:30:53 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-6[59586]: 2026-03-10T05:30:53.404+0000 7f3ec6cd5640 -1 osd.6 38 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail; 57 KiB/s, 0 objects/s recovering 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: osdmap e37: 7 total, 6 up, 7 in 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: from='osd.6 [v2:192.168.123.105:6808/3164767821,v1:192.168.123.105:6809/3164767821]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: Detected new or changed devices on vm05 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: Adjusting osd_memory_target on vm05 to 2176M 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:30:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:53 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:30:54.451 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:54 vm05 ceph-mon[50927]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-10T05:30:54.451 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:54 vm05 ceph-mon[50927]: osdmap e38: 7 total, 6 up, 7 in 2026-03-10T05:30:54.451 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:54 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:54.451 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:54 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:54.451 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:54 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:54.451 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:54 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:54.451 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:54 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:54.451 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:54 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/2230343884' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5e5804cd-52e9-44e4-9407-776a5caf39a6"}]: dispatch 2026-03-10T05:30:54.451 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:54 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/2230343884' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5e5804cd-52e9-44e4-9407-776a5caf39a6"}]': finished 2026-03-10T05:30:54.451 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:54 vm05 ceph-mon[50927]: osd.6 [v2:192.168.123.105:6808/3164767821,v1:192.168.123.105:6809/3164767821] boot 2026-03-10T05:30:54.451 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:54 vm05 ceph-mon[50927]: osdmap e39: 8 total, 7 up, 8 in 2026-03-10T05:30:54.451 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:54 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:54.451 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:54 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:30:54.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:54 vm01 ceph-mon[47941]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-10T05:30:54.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:54 vm01 ceph-mon[47941]: osdmap e38: 7 total, 6 up, 7 in 2026-03-10T05:30:54.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:54 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:54.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:54 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:54.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:54 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:54.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:54 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:54.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:54 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:54.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:54 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/2230343884' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5e5804cd-52e9-44e4-9407-776a5caf39a6"}]: dispatch 2026-03-10T05:30:54.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:54 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/2230343884' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5e5804cd-52e9-44e4-9407-776a5caf39a6"}]': finished 2026-03-10T05:30:54.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:54 vm01 ceph-mon[47941]: osd.6 [v2:192.168.123.105:6808/3164767821,v1:192.168.123.105:6809/3164767821] boot 2026-03-10T05:30:54.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:54 vm01 ceph-mon[47941]: osdmap e39: 8 total, 7 up, 8 in 2026-03-10T05:30:54.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:54 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:54.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:54 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:30:54.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:54 vm02 ceph-mon[50473]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-10T05:30:54.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:54 vm02 ceph-mon[50473]: osdmap e38: 7 total, 6 up, 7 in 2026-03-10T05:30:54.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:54 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:54.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:54 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:54.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:54 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T05:30:54.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:54 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T05:30:54.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:54 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:54.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:54 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/2230343884' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5e5804cd-52e9-44e4-9407-776a5caf39a6"}]: dispatch 2026-03-10T05:30:54.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:54 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/2230343884' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5e5804cd-52e9-44e4-9407-776a5caf39a6"}]': finished 2026-03-10T05:30:54.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:54 vm02 ceph-mon[50473]: osd.6 [v2:192.168.123.105:6808/3164767821,v1:192.168.123.105:6809/3164767821] boot 2026-03-10T05:30:54.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:54 vm02 ceph-mon[50473]: osdmap e39: 8 total, 7 up, 8 in 2026-03-10T05:30:54.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:54 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T05:30:54.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:54 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:30:55.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:55 vm05 ceph-mon[50927]: purged_snaps scrub starts 2026-03-10T05:30:55.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:55 vm05 ceph-mon[50927]: purged_snaps scrub ok 2026-03-10T05:30:55.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:55 vm05 ceph-mon[50927]: from='client.24265 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:55.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:55 vm05 ceph-mon[50927]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail; 57 KiB/s, 0 objects/s recovering 2026-03-10T05:30:55.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:55 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1889891095' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:55.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:55 vm01 ceph-mon[47941]: purged_snaps scrub starts 2026-03-10T05:30:55.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:55 vm01 ceph-mon[47941]: purged_snaps scrub ok 2026-03-10T05:30:55.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:55 vm01 ceph-mon[47941]: from='client.24265 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:55.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:55 vm01 ceph-mon[47941]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail; 57 KiB/s, 0 objects/s recovering 2026-03-10T05:30:55.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:55 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1889891095' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:55.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:55 vm02 ceph-mon[50473]: purged_snaps scrub starts 2026-03-10T05:30:55.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:55 vm02 ceph-mon[50473]: purged_snaps scrub ok 2026-03-10T05:30:55.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:55 vm02 ceph-mon[50473]: from='client.24265 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm05:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:30:55.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:55 vm02 ceph-mon[50473]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail; 57 KiB/s, 0 objects/s recovering 2026-03-10T05:30:55.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:55 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1889891095' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T05:30:56.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:56 vm05 ceph-mon[50927]: osdmap e40: 8 total, 7 up, 8 in 2026-03-10T05:30:56.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:56 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:30:56.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:56 vm01 ceph-mon[47941]: osdmap e40: 8 total, 7 up, 8 in 2026-03-10T05:30:56.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:56 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:30:56.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:56 vm02 ceph-mon[50473]: osdmap e40: 8 total, 7 up, 8 in 2026-03-10T05:30:56.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:56 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:30:57.772 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:57 vm05 ceph-mon[50927]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:30:57.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:57 vm01 ceph-mon[47941]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:30:57.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:57 vm02 ceph-mon[50473]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:30:59.440 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:59 vm05 ceph-mon[50927]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:30:59.440 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:59 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-10T05:30:59.440 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:59 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:59.440 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:30:59 vm05 ceph-mon[50927]: Deploying daemon osd.7 on vm05 2026-03-10T05:30:59.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:59 vm01 ceph-mon[47941]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:30:59.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:59 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-10T05:30:59.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:59 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:59.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:30:59 vm01 ceph-mon[47941]: Deploying daemon osd.7 on vm05 2026-03-10T05:30:59.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:59 vm02 ceph-mon[50473]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:30:59.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:59 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-10T05:30:59.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:59 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:30:59.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:30:59 vm02 ceph-mon[50473]: Deploying daemon osd.7 on vm05 2026-03-10T05:31:01.520 INFO:teuthology.orchestra.run.vm05.stdout:Created osd(s) 7 on host 'vm05' 2026-03-10T05:31:01.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:01 vm05 ceph-mon[50927]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:31:01.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:01 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:01.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:01 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:01 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:01 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:01 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:01 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:01.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:01 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:01.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:01 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.572 DEBUG:teuthology.orchestra.run.vm05:osd.7> sudo journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.7.service 2026-03-10T05:31:01.574 INFO:tasks.cephadm:Waiting for 8 OSDs to come up... 2026-03-10T05:31:01.574 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd stat -f json 2026-03-10T05:31:01.743 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:01.766 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:01 vm01 ceph-mon[47941]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:31:01.766 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:01 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:01.766 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:01 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.767 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:01 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.767 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:01 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.767 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:01 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.767 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:01 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:01.767 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:01 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:01.767 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:01 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:01 vm02 ceph-mon[50473]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:31:01.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:01 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:01.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:01 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:01 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:01 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:01 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:01 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:01.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:01 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:01.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:01 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:01.962 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:31:02.030 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":40,"num_osds":8,"num_up_osds":7,"osd_up_since":1773120654,"num_in_osds":8,"osd_in_since":1773120654,"num_remapped_pgs":0} 2026-03-10T05:31:02.412 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:31:02 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-7[64413]: 2026-03-10T05:31:02.144+0000 7f830f942740 -1 osd.7 0 log_to_monitors true 2026-03-10T05:31:02.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:02 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:02.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:02 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:02.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:02 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:02.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:02 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1713374138' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:31:02.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:02 vm05 ceph-mon[50927]: from='osd.7 [v2:192.168.123.105:6816/2481295119,v1:192.168.123.105:6817/2481295119]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T05:31:02.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:02 vm05 ceph-mon[50927]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T05:31:02.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:02 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:02.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:02 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:02.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:02 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:02.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:02 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1713374138' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:31:02.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:02 vm01 ceph-mon[47941]: from='osd.7 [v2:192.168.123.105:6816/2481295119,v1:192.168.123.105:6817/2481295119]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T05:31:02.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:02 vm01 ceph-mon[47941]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T05:31:02.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:02 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:02.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:02 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:02.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:02 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:02.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:02 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1713374138' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:31:02.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:02 vm02 ceph-mon[50473]: from='osd.7 [v2:192.168.123.105:6816/2481295119,v1:192.168.123.105:6817/2481295119]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T05:31:02.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:02 vm02 ceph-mon[50473]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T05:31:03.031 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd stat -f json 2026-03-10T05:31:03.196 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:03.424 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:31:03.488 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":41,"num_osds":8,"num_up_osds":7,"osd_up_since":1773120654,"num_in_osds":8,"osd_in_since":1773120654,"num_remapped_pgs":0} 2026-03-10T05:31:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:31:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: Detected new or changed devices on vm05 2026-03-10T05:31:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:31:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:31:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:31:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: Adjusting osd_memory_target on vm05 to 1451M 2026-03-10T05:31:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-10T05:31:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: osdmap e41: 8 total, 7 up, 8 in 2026-03-10T05:31:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:31:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: from='osd.7 [v2:192.168.123.105:6816/2481295119,v1:192.168.123.105:6817/2481295119]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:31:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:31:03.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:03 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2885845371' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:31:03.805 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:31:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-7[64413]: 2026-03-10T05:31:03.529+0000 7f830c0d6640 -1 osd.7 0 waiting for initial osdmap 2026-03-10T05:31:03.805 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:31:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-7[64413]: 2026-03-10T05:31:03.538+0000 7f83076ed640 -1 osd.7 42 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: Detected new or changed devices on vm05 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: Adjusting osd_memory_target on vm05 to 1451M 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: osdmap e41: 8 total, 7 up, 8 in 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: from='osd.7 [v2:192.168.123.105:6816/2481295119,v1:192.168.123.105:6817/2481295119]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:31:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:03 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2885845371' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:31:03.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:31:03.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: Detected new or changed devices on vm05 2026-03-10T05:31:03.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:03.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:03.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:31:03.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:31:03.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"}]: dispatch 2026-03-10T05:31:03.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: Adjusting osd_memory_target on vm05 to 1451M 2026-03-10T05:31:03.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:03.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:03.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:03.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:03.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-10T05:31:03.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: osdmap e41: 8 total, 7 up, 8 in 2026-03-10T05:31:03.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:31:03.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: from='osd.7 [v2:192.168.123.105:6816/2481295119,v1:192.168.123.105:6817/2481295119]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:31:03.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]: dispatch 2026-03-10T05:31:03.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:03 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2885845371' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:31:04.490 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd stat -f json 2026-03-10T05:31:04.668 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:04.785 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:04 vm01 ceph-mon[47941]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-10T05:31:04.785 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:04 vm01 ceph-mon[47941]: osdmap e42: 8 total, 7 up, 8 in 2026-03-10T05:31:04.785 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:04 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:31:04.785 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:04 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:31:04.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:04 vm05 ceph-mon[50927]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-10T05:31:04.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:04 vm05 ceph-mon[50927]: osdmap e42: 8 total, 7 up, 8 in 2026-03-10T05:31:04.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:04 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:31:04.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:04 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:31:04.892 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:31:04.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:04 vm02 ceph-mon[50473]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm05", "root=default"]}]': finished 2026-03-10T05:31:04.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:04 vm02 ceph-mon[50473]: osdmap e42: 8 total, 7 up, 8 in 2026-03-10T05:31:04.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:04 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:31:04.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:04 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:31:04.961 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":43,"num_osds":8,"num_up_osds":8,"osd_up_since":1773120664,"num_in_osds":8,"osd_in_since":1773120654,"num_remapped_pgs":0} 2026-03-10T05:31:04.961 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd dump --format=json 2026-03-10T05:31:05.131 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:05.370 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:31:05.371 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":43,"fsid":"dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba","created":"2026-03-10T05:28:48.469911+0000","modified":"2026-03-10T05:31:04.524556+0000","last_up_change":"2026-03-10T05:31:04.524556+0000","last_in_change":"2026-03-10T05:30:54.334085+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T05:30:15.890898+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"22","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"e9a77b3b-eb3e-404d-816d-a374f1e03f47","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":25,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":2152256884},{"type":"v1","addr":"192.168.123.101:6803","nonce":2152256884}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":2152256884},{"type":"v1","addr":"192.168.123.101:6805","nonce":2152256884}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":2152256884},{"type":"v1","addr":"192.168.123.101:6809","nonce":2152256884}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":2152256884},{"type":"v1","addr":"192.168.123.101:6807","nonce":2152256884}]},"public_addr":"192.168.123.101:6803/2152256884","cluster_addr":"192.168.123.101:6805/2152256884","heartbeat_back_addr":"192.168.123.101:6809/2152256884","heartbeat_front_addr":"192.168.123.101:6807/2152256884","state":["exists","up"]},{"osd":1,"uuid":"423e15ff-ce61-4730-b91e-0f9ca81dd61d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":14,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":1960293858},{"type":"v1","addr":"192.168.123.101:6811","nonce":1960293858}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":1960293858},{"type":"v1","addr":"192.168.123.101:6813","nonce":1960293858}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":1960293858},{"type":"v1","addr":"192.168.123.101:6817","nonce":1960293858}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":1960293858},{"type":"v1","addr":"192.168.123.101:6815","nonce":1960293858}]},"public_addr":"192.168.123.101:6811/1960293858","cluster_addr":"192.168.123.101:6813/1960293858","heartbeat_back_addr":"192.168.123.101:6817/1960293858","heartbeat_front_addr":"192.168.123.101:6815/1960293858","state":["exists","up"]},{"osd":2,"uuid":"e19eb7fe-8764-496e-ad96-eed37a0f93ad","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":20,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6800","nonce":1562513062},{"type":"v1","addr":"192.168.123.102:6801","nonce":1562513062}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":1562513062},{"type":"v1","addr":"192.168.123.102:6803","nonce":1562513062}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":1562513062},{"type":"v1","addr":"192.168.123.102:6807","nonce":1562513062}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":1562513062},{"type":"v1","addr":"192.168.123.102:6805","nonce":1562513062}]},"public_addr":"192.168.123.102:6801/1562513062","cluster_addr":"192.168.123.102:6803/1562513062","heartbeat_back_addr":"192.168.123.102:6807/1562513062","heartbeat_front_addr":"192.168.123.102:6805/1562513062","state":["exists","up"]},{"osd":3,"uuid":"9916ed76-ed97-49db-960f-84cec5746a36","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":35,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":914771124},{"type":"v1","addr":"192.168.123.102:6809","nonce":914771124}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6810","nonce":914771124},{"type":"v1","addr":"192.168.123.102:6811","nonce":914771124}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6814","nonce":914771124},{"type":"v1","addr":"192.168.123.102:6815","nonce":914771124}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":914771124},{"type":"v1","addr":"192.168.123.102:6813","nonce":914771124}]},"public_addr":"192.168.123.102:6809/914771124","cluster_addr":"192.168.123.102:6811/914771124","heartbeat_back_addr":"192.168.123.102:6815/914771124","heartbeat_front_addr":"192.168.123.102:6813/914771124","state":["exists","up"]},{"osd":4,"uuid":"cb979506-99a8-4950-b248-04eeab920b11","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":30,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":4203989344},{"type":"v1","addr":"192.168.123.102:6817","nonce":4203989344}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6818","nonce":4203989344},{"type":"v1","addr":"192.168.123.102:6819","nonce":4203989344}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6822","nonce":4203989344},{"type":"v1","addr":"192.168.123.102:6823","nonce":4203989344}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6820","nonce":4203989344},{"type":"v1","addr":"192.168.123.102:6821","nonce":4203989344}]},"public_addr":"192.168.123.102:6817/4203989344","cluster_addr":"192.168.123.102:6819/4203989344","heartbeat_back_addr":"192.168.123.102:6823/4203989344","heartbeat_front_addr":"192.168.123.102:6821/4203989344","state":["exists","up"]},{"osd":5,"uuid":"4615618e-f690-4722-b0c7-d3d66c66ee61","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":34,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":576190648},{"type":"v1","addr":"192.168.123.105:6801","nonce":576190648}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":576190648},{"type":"v1","addr":"192.168.123.105:6803","nonce":576190648}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":576190648},{"type":"v1","addr":"192.168.123.105:6807","nonce":576190648}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":576190648},{"type":"v1","addr":"192.168.123.105:6805","nonce":576190648}]},"public_addr":"192.168.123.105:6801/576190648","cluster_addr":"192.168.123.105:6803/576190648","heartbeat_back_addr":"192.168.123.105:6807/576190648","heartbeat_front_addr":"192.168.123.105:6805/576190648","state":["exists","up"]},{"osd":6,"uuid":"33f0f20d-1d48-4480-9474-610fa90103f6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":39,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6808","nonce":3164767821},{"type":"v1","addr":"192.168.123.105:6809","nonce":3164767821}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6810","nonce":3164767821},{"type":"v1","addr":"192.168.123.105:6811","nonce":3164767821}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6814","nonce":3164767821},{"type":"v1","addr":"192.168.123.105:6815","nonce":3164767821}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6812","nonce":3164767821},{"type":"v1","addr":"192.168.123.105:6813","nonce":3164767821}]},"public_addr":"192.168.123.105:6809/3164767821","cluster_addr":"192.168.123.105:6811/3164767821","heartbeat_back_addr":"192.168.123.105:6815/3164767821","heartbeat_front_addr":"192.168.123.105:6813/3164767821","state":["exists","up"]},{"osd":7,"uuid":"5e5804cd-52e9-44e4-9407-776a5caf39a6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":43,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6816","nonce":2481295119},{"type":"v1","addr":"192.168.123.105:6817","nonce":2481295119}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6818","nonce":2481295119},{"type":"v1","addr":"192.168.123.105:6819","nonce":2481295119}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6822","nonce":2481295119},{"type":"v1","addr":"192.168.123.105:6823","nonce":2481295119}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6820","nonce":2481295119},{"type":"v1","addr":"192.168.123.105:6821","nonce":2481295119}]},"public_addr":"192.168.123.105:6817/2481295119","cluster_addr":"192.168.123.105:6819/2481295119","heartbeat_back_addr":"192.168.123.105:6823/2481295119","heartbeat_front_addr":"192.168.123.105:6821/2481295119","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:29:51.081158+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:01.736940+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:11.825689+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:22.626076+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:33.635533+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:42.045408+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:52.918124+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:0/1029535926":"2026-03-11T05:29:09.828614+0000","192.168.123.101:0/1488281479":"2026-03-11T05:29:09.828614+0000","192.168.123.101:6801/798997338":"2026-03-11T05:29:09.828614+0000","192.168.123.101:0/2671353675":"2026-03-11T05:29:09.828614+0000","192.168.123.101:0/1596225986":"2026-03-11T05:28:59.352838+0000","192.168.123.101:0/1037707188":"2026-03-11T05:28:59.352838+0000","192.168.123.101:6800/798997338":"2026-03-11T05:29:09.828614+0000","192.168.123.101:0/2614637726":"2026-03-11T05:28:59.352838+0000","192.168.123.101:6801/1652942659":"2026-03-11T05:28:59.352838+0000","192.168.123.101:6800/1652942659":"2026-03-11T05:28:59.352838+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T05:31:05.439 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-10T05:30:15.890898+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '22', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 7.889999866485596, 'score_stable': 7.889999866485596, 'optimal_score': 0.3799999952316284, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-10T05:31:05.439 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd pool get .mgr pg_num 2026-03-10T05:31:05.612 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:05.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:05 vm01 ceph-mon[47941]: purged_snaps scrub starts 2026-03-10T05:31:05.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:05 vm01 ceph-mon[47941]: purged_snaps scrub ok 2026-03-10T05:31:05.665 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:05 vm01 ceph-mon[47941]: pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:31:05.666 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:05 vm01 ceph-mon[47941]: osd.7 [v2:192.168.123.105:6816/2481295119,v1:192.168.123.105:6817/2481295119] boot 2026-03-10T05:31:05.666 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:05 vm01 ceph-mon[47941]: osdmap e43: 8 total, 8 up, 8 in 2026-03-10T05:31:05.666 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:05 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:31:05.666 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:05 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3922691261' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:31:05.666 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:05 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4285353164' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:31:05.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:05 vm05 ceph-mon[50927]: purged_snaps scrub starts 2026-03-10T05:31:05.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:05 vm05 ceph-mon[50927]: purged_snaps scrub ok 2026-03-10T05:31:05.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:05 vm05 ceph-mon[50927]: pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:31:05.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:05 vm05 ceph-mon[50927]: osd.7 [v2:192.168.123.105:6816/2481295119,v1:192.168.123.105:6817/2481295119] boot 2026-03-10T05:31:05.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:05 vm05 ceph-mon[50927]: osdmap e43: 8 total, 8 up, 8 in 2026-03-10T05:31:05.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:05 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:31:05.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:05 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3922691261' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:31:05.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:05 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4285353164' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:31:05.836 INFO:teuthology.orchestra.run.vm01.stdout:pg_num: 1 2026-03-10T05:31:05.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:05 vm02 ceph-mon[50473]: purged_snaps scrub starts 2026-03-10T05:31:05.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:05 vm02 ceph-mon[50473]: purged_snaps scrub ok 2026-03-10T05:31:05.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:05 vm02 ceph-mon[50473]: pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 187 MiB used, 140 GiB / 140 GiB avail 2026-03-10T05:31:05.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:05 vm02 ceph-mon[50473]: osd.7 [v2:192.168.123.105:6816/2481295119,v1:192.168.123.105:6817/2481295119] boot 2026-03-10T05:31:05.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:05 vm02 ceph-mon[50473]: osdmap e43: 8 total, 8 up, 8 in 2026-03-10T05:31:05.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:05 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T05:31:05.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:05 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3922691261' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T05:31:05.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:05 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4285353164' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:31:05.908 INFO:tasks.cephadm:Adding ceph.iscsi.iscsi.a on vm01 2026-03-10T05:31:05.908 INFO:tasks.cephadm:Adding ceph.iscsi.iscsi.b on vm05 2026-03-10T05:31:05.909 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd pool create datapool 3 3 replicated 2026-03-10T05:31:06.076 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.c/config 2026-03-10T05:31:06.551 INFO:teuthology.orchestra.run.vm05.stderr:pool 'datapool' created 2026-03-10T05:31:06.606 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- rbd pool init datapool 2026-03-10T05:31:06.769 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.c/config 2026-03-10T05:31:06.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:06 vm05 ceph-mon[50927]: osdmap e44: 8 total, 8 up, 8 in 2026-03-10T05:31:06.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:06 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/466454476' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T05:31:06.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:06 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1289566113' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-10T05:31:06.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:06 vm05 ceph-mon[50927]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-10T05:31:06.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:06 vm01 ceph-mon[47941]: osdmap e44: 8 total, 8 up, 8 in 2026-03-10T05:31:06.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:06 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/466454476' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T05:31:06.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:06 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1289566113' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-10T05:31:06.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:06 vm01 ceph-mon[47941]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-10T05:31:06.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:06 vm02 ceph-mon[50473]: osdmap e44: 8 total, 8 up, 8 in 2026-03-10T05:31:06.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:06 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/466454476' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T05:31:06.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:06 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1289566113' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-10T05:31:06.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:06 vm02 ceph-mon[50473]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]: dispatch 2026-03-10T05:31:07.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:07 vm05 ceph-mon[50927]: pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:31:07.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:07 vm05 ceph-mon[50927]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-10T05:31:07.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:07 vm05 ceph-mon[50927]: osdmap e45: 8 total, 8 up, 8 in 2026-03-10T05:31:07.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:07 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3730712291' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-10T05:31:07.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:07 vm05 ceph-mon[50927]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-10T05:31:07.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:07 vm01 ceph-mon[47941]: pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:31:07.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:07 vm01 ceph-mon[47941]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-10T05:31:07.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:07 vm01 ceph-mon[47941]: osdmap e45: 8 total, 8 up, 8 in 2026-03-10T05:31:07.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:07 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3730712291' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-10T05:31:07.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:07 vm01 ceph-mon[47941]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-10T05:31:07.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:07 vm02 ceph-mon[50473]: pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:31:07.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:07 vm02 ceph-mon[50473]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-10T05:31:07.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:07 vm02 ceph-mon[50473]: osdmap e45: 8 total, 8 up, 8 in 2026-03-10T05:31:07.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:07 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3730712291' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-10T05:31:07.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:07 vm02 ceph-mon[50473]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]: dispatch 2026-03-10T05:31:08.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:08 vm01 ceph-mon[47941]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-10T05:31:08.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:08 vm01 ceph-mon[47941]: osdmap e46: 8 total, 8 up, 8 in 2026-03-10T05:31:08.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:08 vm01 ceph-mon[47941]: pgmap v95: 4 pgs: 3 unknown, 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:31:08.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:08 vm02 ceph-mon[50473]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-10T05:31:08.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:08 vm02 ceph-mon[50473]: osdmap e46: 8 total, 8 up, 8 in 2026-03-10T05:31:08.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:08 vm02 ceph-mon[50473]: pgmap v95: 4 pgs: 3 unknown, 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:31:09.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:08 vm05 ceph-mon[50927]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-10T05:31:09.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:08 vm05 ceph-mon[50927]: osdmap e46: 8 total, 8 up, 8 in 2026-03-10T05:31:09.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:08 vm05 ceph-mon[50927]: pgmap v95: 4 pgs: 3 unknown, 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T05:31:09.618 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph orch apply iscsi datapool admin admin --trusted_ip_list 192.168.123.101,192.168.123.105 --placement '2;vm01=iscsi.a;vm05=iscsi.b' 2026-03-10T05:31:09.791 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.c/config 2026-03-10T05:31:09.832 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:09 vm05 ceph-mon[50927]: osdmap e47: 8 total, 8 up, 8 in 2026-03-10T05:31:09.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:09 vm01 ceph-mon[47941]: osdmap e47: 8 total, 8 up, 8 in 2026-03-10T05:31:09.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:09 vm02 ceph-mon[50473]: osdmap e47: 8 total, 8 up, 8 in 2026-03-10T05:31:10.037 INFO:teuthology.orchestra.run.vm05.stdout:Scheduled iscsi.datapool update... 2026-03-10T05:31:10.098 INFO:tasks.cephadm:Distributing iscsi-gateway.cfg... 2026-03-10T05:31:10.098 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:31:10.098 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/iscsi-gateway.cfg 2026-03-10T05:31:10.132 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:31:10.132 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/iscsi-gateway.cfg 2026-03-10T05:31:10.158 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T05:31:10.158 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/ceph/iscsi-gateway.cfg 2026-03-10T05:31:10.186 DEBUG:teuthology.orchestra.run.vm01:iscsi.iscsi.a> sudo journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@iscsi.iscsi.a.service 2026-03-10T05:31:10.187 DEBUG:teuthology.orchestra.run.vm05:iscsi.iscsi.b> sudo journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@iscsi.iscsi.b.service 2026-03-10T05:31:10.227 INFO:tasks.cephadm:Setting up client nodes... 2026-03-10T05:31:10.227 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph auth get-or-create client.0 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-10T05:31:10.487 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:10.823 INFO:teuthology.orchestra.run.vm01.stdout:[client.0] 2026-03-10T05:31:10.823 INFO:teuthology.orchestra.run.vm01.stdout: key = AQCerK9p/cGRMBAAEfWB6Ky++OBOEirpPfIUrg== 2026-03-10T05:31:10.823 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:10 vm01 systemd[1]: Starting Ceph iscsi.iscsi.a for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:31:10.823 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:10 vm01 ceph-mon[47941]: osdmap e48: 8 total, 8 up, 8 in 2026-03-10T05:31:10.823 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:10 vm01 ceph-mon[47941]: pgmap v98: 4 pgs: 3 active+clean, 1 unknown; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 511 B/s wr, 0 op/s 2026-03-10T05:31:10.823 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:10 vm01 ceph-mon[47941]: from='client.14424 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.101,192.168.123.105", "placement": "2;vm01=iscsi.a;vm05=iscsi.b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:10.823 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:10 vm01 ceph-mon[47941]: Saving service iscsi.datapool spec with placement vm01=iscsi.a;vm05=iscsi.b;count:2 2026-03-10T05:31:10.823 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:10.823 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:10.823 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:10.823 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:10.823 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:10.823 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-10T05:31:10.824 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-10T05:31:10.824 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:10 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:10.824 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:10 vm01 ceph-mon[47941]: Deploying daemon iscsi.iscsi.a on vm01 2026-03-10T05:31:10.886 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:31:10.886 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.client.0.keyring 2026-03-10T05:31:10.886 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod 0644 /etc/ceph/ceph.client.0.keyring 2026-03-10T05:31:10.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:10 vm02 ceph-mon[50473]: osdmap e48: 8 total, 8 up, 8 in 2026-03-10T05:31:10.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:10 vm02 ceph-mon[50473]: pgmap v98: 4 pgs: 3 active+clean, 1 unknown; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 511 B/s wr, 0 op/s 2026-03-10T05:31:10.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:10 vm02 ceph-mon[50473]: from='client.14424 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.101,192.168.123.105", "placement": "2;vm01=iscsi.a;vm05=iscsi.b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:10.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:10 vm02 ceph-mon[50473]: Saving service iscsi.datapool spec with placement vm01=iscsi.a;vm05=iscsi.b;count:2 2026-03-10T05:31:10.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:10.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:10.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:10.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:10.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:10.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-10T05:31:10.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-10T05:31:10.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:10 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:10.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:10 vm02 ceph-mon[50473]: Deploying daemon iscsi.iscsi.a on vm01 2026-03-10T05:31:10.932 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph auth get-or-create client.1 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-10T05:31:11.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:10 vm05 ceph-mon[50927]: osdmap e48: 8 total, 8 up, 8 in 2026-03-10T05:31:11.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:10 vm05 ceph-mon[50927]: pgmap v98: 4 pgs: 3 active+clean, 1 unknown; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 511 B/s wr, 0 op/s 2026-03-10T05:31:11.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:10 vm05 ceph-mon[50927]: from='client.14424 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.101,192.168.123.105", "placement": "2;vm01=iscsi.a;vm05=iscsi.b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:11.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:10 vm05 ceph-mon[50927]: Saving service iscsi.datapool spec with placement vm01=iscsi.a;vm05=iscsi.b;count:2 2026-03-10T05:31:11.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:11.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:11.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:11.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:11.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:11.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-10T05:31:11.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-10T05:31:11.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:10 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:11.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:10 vm05 ceph-mon[50927]: Deploying daemon iscsi.iscsi.a on vm01 2026-03-10T05:31:11.090 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.b/config 2026-03-10T05:31:11.096 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:10 vm01 podman[66461]: 2026-03-10 05:31:10.914647739 +0000 UTC m=+0.017665453 container create 87eedcfb11be985407e933412b07619f5f228f5d72e96240fa07cd7a9315a318 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, ceph=True, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223) 2026-03-10T05:31:11.096 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:10 vm01 podman[66461]: 2026-03-10 05:31:10.94956261 +0000 UTC m=+0.052580324 container init 87eedcfb11be985407e933412b07619f5f228f5d72e96240fa07cd7a9315a318 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, OSD_FLAVOR=default, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T05:31:11.096 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:10 vm01 podman[66461]: 2026-03-10 05:31:10.952237106 +0000 UTC m=+0.055254820 container start 87eedcfb11be985407e933412b07619f5f228f5d72e96240fa07cd7a9315a318 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a, io.buildah.version=1.41.3, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, ceph=True, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-10T05:31:11.096 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:10 vm01 bash[66461]: 87eedcfb11be985407e933412b07619f5f228f5d72e96240fa07cd7a9315a318 2026-03-10T05:31:11.096 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:10 vm01 podman[66461]: 2026-03-10 05:31:10.906828382 +0000 UTC m=+0.009846106 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T05:31:11.096 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:10 vm01 systemd[1]: Started Ceph iscsi.iscsi.a for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:31:11.356 INFO:teuthology.orchestra.run.vm02.stdout:[client.1] 2026-03-10T05:31:11.356 INFO:teuthology.orchestra.run.vm02.stdout: key = AQCfrK9pMQHSFBAAc4W5LNMSxc/HnBCvOfK+KQ== 2026-03-10T05:31:11.373 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Started the configuration object watcher 2026-03-10T05:31:11.373 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Processing osd blocklist entries for this node 2026-03-10T05:31:11.373 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Checking for config object changes every 1s 2026-03-10T05:31:11.373 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Removing blocklisted entry for this host : 192.168.123.101:0/1029535926 2026-03-10T05:31:11.422 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:31:11.422 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.client.1.keyring 2026-03-10T05:31:11.422 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 0644 /etc/ceph/ceph.client.1.keyring 2026-03-10T05:31:11.460 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph auth get-or-create client.2 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-10T05:31:11.674 INFO:teuthology.orchestra.run.vm05.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.c/config 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/933362678' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: Deploying daemon iscsi.iscsi.b on vm05 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: from='client.? 192.168.123.102:0/3810848775' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3003068267' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-10T05:31:11.749 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:11 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2125719779' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1029535926"}]: dispatch 2026-03-10T05:31:11.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/933362678' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:11.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:11.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T05:31:11.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:11.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:11.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:11.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-10T05:31:11.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-10T05:31:11.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:11.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: Deploying daemon iscsi.iscsi.b on vm05 2026-03-10T05:31:11.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: from='client.? 192.168.123.102:0/3810848775' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:11.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:11.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T05:31:11.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3003068267' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-10T05:31:11.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:11 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2125719779' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1029535926"}]: dispatch 2026-03-10T05:31:11.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/933362678' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T05:31:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]: dispatch 2026-03-10T05:31:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-10T05:31:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: Deploying daemon iscsi.iscsi.b on vm05 2026-03-10T05:31:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: from='client.? 192.168.123.102:0/3810848775' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T05:31:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3003068267' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-10T05:31:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:11 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2125719779' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1029535926"}]: dispatch 2026-03-10T05:31:12.054 INFO:teuthology.orchestra.run.vm05.stdout:[client.2] 2026-03-10T05:31:12.054 INFO:teuthology.orchestra.run.vm05.stdout: key = AQCgrK9pKenQARAA6zdTe9v9PzVPZCCQ2i602A== 2026-03-10T05:31:12.175 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T05:31:12.175 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/ceph/ceph.client.2.keyring 2026-03-10T05:31:12.175 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod 0644 /etc/ceph/ceph.client.2.keyring 2026-03-10T05:31:12.281 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-10T05:31:12.281 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-10T05:31:12.281 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph mgr dump --format=json 2026-03-10T05:31:12.531 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:12.563 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Successfully removed blocklist entry 2026-03-10T05:31:12.563 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Removing blocklisted entry for this host : 192.168.123.101:0/1488281479 2026-03-10T05:31:12.839 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:31:12.901 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":13,"flags":0,"active_gid":14150,"active_name":"x","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6800","nonce":987764607},{"type":"v1","addr":"192.168.123.101:6801","nonce":987764607}]},"active_addr":"192.168.123.101:6801/987764607","active_change":"2026-03-10T05:29:09.828701+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.101:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":3807485068}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":417658875}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":3406416254}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.101:0","nonce":3998765943}]}]} 2026-03-10T05:31:12.902 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-10T05:31:12.902 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-10T05:31:12.902 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd dump --format=json 2026-03-10T05:31:13.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: pgmap v99: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 385 B/s wr, 0 op/s 2026-03-10T05:31:13.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: Checking pool "datapool" exists for service iscsi.datapool 2026-03-10T05:31:13.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:13.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/622682481' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/622682481' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2125719779' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1029535926"}]': finished 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: osdmap e49: 8 total, 8 up, 8 in 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3712128055' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3510290556' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1488281479"}]: dispatch 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3636762476' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm01"}]: dispatch 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm05"}]: dispatch 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:13.056 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:12 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:13.076 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: pgmap v99: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 385 B/s wr, 0 op/s 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: Checking pool "datapool" exists for service iscsi.datapool 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/622682481' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/622682481' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2125719779' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1029535926"}]': finished 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: osdmap e49: 8 total, 8 up, 8 in 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3712128055' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3510290556' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1488281479"}]: dispatch 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:13.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3636762476' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T05:31:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-10T05:31:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-10T05:31:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm01"}]: dispatch 2026-03-10T05:31:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm05"}]: dispatch 2026-03-10T05:31:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:13.089 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:12 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:13.299 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:31:13.299 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":49,"fsid":"dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba","created":"2026-03-10T05:28:48.469911+0000","modified":"2026-03-10T05:31:12.351396+0000","last_up_change":"2026-03-10T05:31:04.524556+0000","last_in_change":"2026-03-10T05:30:54.334085+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":2,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T05:30:15.890898+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"22","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":"datapool","create_time":"2026-03-10T05:31:06.316398+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":3,"pg_placement_num":3,"pg_placement_num_target":3,"pg_num_target":3,"pg_num_pending":3,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"48","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":48,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":5.3299999237060547,"score_stable":5.3299999237060547,"optimal_score":0.75,"raw_score_acting":4,"raw_score_stable":4,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"e9a77b3b-eb3e-404d-816d-a374f1e03f47","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":25,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":2152256884},{"type":"v1","addr":"192.168.123.101:6803","nonce":2152256884}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":2152256884},{"type":"v1","addr":"192.168.123.101:6805","nonce":2152256884}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":2152256884},{"type":"v1","addr":"192.168.123.101:6809","nonce":2152256884}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":2152256884},{"type":"v1","addr":"192.168.123.101:6807","nonce":2152256884}]},"public_addr":"192.168.123.101:6803/2152256884","cluster_addr":"192.168.123.101:6805/2152256884","heartbeat_back_addr":"192.168.123.101:6809/2152256884","heartbeat_front_addr":"192.168.123.101:6807/2152256884","state":["exists","up"]},{"osd":1,"uuid":"423e15ff-ce61-4730-b91e-0f9ca81dd61d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":14,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":1960293858},{"type":"v1","addr":"192.168.123.101:6811","nonce":1960293858}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":1960293858},{"type":"v1","addr":"192.168.123.101:6813","nonce":1960293858}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":1960293858},{"type":"v1","addr":"192.168.123.101:6817","nonce":1960293858}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":1960293858},{"type":"v1","addr":"192.168.123.101:6815","nonce":1960293858}]},"public_addr":"192.168.123.101:6811/1960293858","cluster_addr":"192.168.123.101:6813/1960293858","heartbeat_back_addr":"192.168.123.101:6817/1960293858","heartbeat_front_addr":"192.168.123.101:6815/1960293858","state":["exists","up"]},{"osd":2,"uuid":"e19eb7fe-8764-496e-ad96-eed37a0f93ad","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":45,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6800","nonce":1562513062},{"type":"v1","addr":"192.168.123.102:6801","nonce":1562513062}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":1562513062},{"type":"v1","addr":"192.168.123.102:6803","nonce":1562513062}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":1562513062},{"type":"v1","addr":"192.168.123.102:6807","nonce":1562513062}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":1562513062},{"type":"v1","addr":"192.168.123.102:6805","nonce":1562513062}]},"public_addr":"192.168.123.102:6801/1562513062","cluster_addr":"192.168.123.102:6803/1562513062","heartbeat_back_addr":"192.168.123.102:6807/1562513062","heartbeat_front_addr":"192.168.123.102:6805/1562513062","state":["exists","up"]},{"osd":3,"uuid":"9916ed76-ed97-49db-960f-84cec5746a36","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":45,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":914771124},{"type":"v1","addr":"192.168.123.102:6809","nonce":914771124}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6810","nonce":914771124},{"type":"v1","addr":"192.168.123.102:6811","nonce":914771124}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6814","nonce":914771124},{"type":"v1","addr":"192.168.123.102:6815","nonce":914771124}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":914771124},{"type":"v1","addr":"192.168.123.102:6813","nonce":914771124}]},"public_addr":"192.168.123.102:6809/914771124","cluster_addr":"192.168.123.102:6811/914771124","heartbeat_back_addr":"192.168.123.102:6815/914771124","heartbeat_front_addr":"192.168.123.102:6813/914771124","state":["exists","up"]},{"osd":4,"uuid":"cb979506-99a8-4950-b248-04eeab920b11","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":30,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":4203989344},{"type":"v1","addr":"192.168.123.102:6817","nonce":4203989344}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6818","nonce":4203989344},{"type":"v1","addr":"192.168.123.102:6819","nonce":4203989344}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6822","nonce":4203989344},{"type":"v1","addr":"192.168.123.102:6823","nonce":4203989344}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6820","nonce":4203989344},{"type":"v1","addr":"192.168.123.102:6821","nonce":4203989344}]},"public_addr":"192.168.123.102:6817/4203989344","cluster_addr":"192.168.123.102:6819/4203989344","heartbeat_back_addr":"192.168.123.102:6823/4203989344","heartbeat_front_addr":"192.168.123.102:6821/4203989344","state":["exists","up"]},{"osd":5,"uuid":"4615618e-f690-4722-b0c7-d3d66c66ee61","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":34,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":576190648},{"type":"v1","addr":"192.168.123.105:6801","nonce":576190648}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":576190648},{"type":"v1","addr":"192.168.123.105:6803","nonce":576190648}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":576190648},{"type":"v1","addr":"192.168.123.105:6807","nonce":576190648}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":576190648},{"type":"v1","addr":"192.168.123.105:6805","nonce":576190648}]},"public_addr":"192.168.123.105:6801/576190648","cluster_addr":"192.168.123.105:6803/576190648","heartbeat_back_addr":"192.168.123.105:6807/576190648","heartbeat_front_addr":"192.168.123.105:6805/576190648","state":["exists","up"]},{"osd":6,"uuid":"33f0f20d-1d48-4480-9474-610fa90103f6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":39,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6808","nonce":3164767821},{"type":"v1","addr":"192.168.123.105:6809","nonce":3164767821}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6810","nonce":3164767821},{"type":"v1","addr":"192.168.123.105:6811","nonce":3164767821}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6814","nonce":3164767821},{"type":"v1","addr":"192.168.123.105:6815","nonce":3164767821}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6812","nonce":3164767821},{"type":"v1","addr":"192.168.123.105:6813","nonce":3164767821}]},"public_addr":"192.168.123.105:6809/3164767821","cluster_addr":"192.168.123.105:6811/3164767821","heartbeat_back_addr":"192.168.123.105:6815/3164767821","heartbeat_front_addr":"192.168.123.105:6813/3164767821","state":["exists","up"]},{"osd":7,"uuid":"5e5804cd-52e9-44e4-9407-776a5caf39a6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":43,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6816","nonce":2481295119},{"type":"v1","addr":"192.168.123.105:6817","nonce":2481295119}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6818","nonce":2481295119},{"type":"v1","addr":"192.168.123.105:6819","nonce":2481295119}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6822","nonce":2481295119},{"type":"v1","addr":"192.168.123.105:6823","nonce":2481295119}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6820","nonce":2481295119},{"type":"v1","addr":"192.168.123.105:6821","nonce":2481295119}]},"public_addr":"192.168.123.105:6817/2481295119","cluster_addr":"192.168.123.105:6819/2481295119","heartbeat_back_addr":"192.168.123.105:6823/2481295119","heartbeat_front_addr":"192.168.123.105:6821/2481295119","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:29:51.081158+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:01.736940+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:11.825689+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:22.626076+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:33.635533+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:42.045408+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:52.918124+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:31:03.110766+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:0/1488281479":"2026-03-11T05:29:09.828614+0000","192.168.123.101:6801/798997338":"2026-03-11T05:29:09.828614+0000","192.168.123.101:0/2671353675":"2026-03-11T05:29:09.828614+0000","192.168.123.101:0/1596225986":"2026-03-11T05:28:59.352838+0000","192.168.123.101:0/1037707188":"2026-03-11T05:28:59.352838+0000","192.168.123.101:6800/798997338":"2026-03-11T05:29:09.828614+0000","192.168.123.101:0/2614637726":"2026-03-11T05:28:59.352838+0000","192.168.123.101:6801/1652942659":"2026-03-11T05:28:59.352838+0000","192.168.123.101:6800/1652942659":"2026-03-11T05:28:59.352838+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T05:31:13.365 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-10T05:31:13.365 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd dump --format=json 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: pgmap v99: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 385 B/s wr, 0 op/s 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: Checking pool "datapool" exists for service iscsi.datapool 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/622682481' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/622682481' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2125719779' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1029535926"}]': finished 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: osdmap e49: 8 total, 8 up, 8 in 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3712128055' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist ls"}]: dispatch 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3510290556' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1488281479"}]: dispatch 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3636762476' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm01"}]: dispatch 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm05"}]: dispatch 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:31:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:12 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:31:13.556 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:13.615 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Successfully removed blocklist entry 2026-03-10T05:31:13.615 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Removing blocklisted entry for this host : 192.168.123.101:6801/798997338 2026-03-10T05:31:13.783 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:31:13.783 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":50,"fsid":"dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba","created":"2026-03-10T05:28:48.469911+0000","modified":"2026-03-10T05:31:13.354235+0000","last_up_change":"2026-03-10T05:31:04.524556+0000","last_in_change":"2026-03-10T05:30:54.334085+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":2,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T05:30:15.890898+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"22","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":"datapool","create_time":"2026-03-10T05:31:06.316398+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":3,"pg_placement_num":3,"pg_placement_num_target":3,"pg_num_target":3,"pg_num_pending":3,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"48","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":48,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":5.3299999237060547,"score_stable":5.3299999237060547,"optimal_score":0.75,"raw_score_acting":4,"raw_score_stable":4,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"e9a77b3b-eb3e-404d-816d-a374f1e03f47","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":25,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":2152256884},{"type":"v1","addr":"192.168.123.101:6803","nonce":2152256884}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":2152256884},{"type":"v1","addr":"192.168.123.101:6805","nonce":2152256884}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6808","nonce":2152256884},{"type":"v1","addr":"192.168.123.101:6809","nonce":2152256884}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":2152256884},{"type":"v1","addr":"192.168.123.101:6807","nonce":2152256884}]},"public_addr":"192.168.123.101:6803/2152256884","cluster_addr":"192.168.123.101:6805/2152256884","heartbeat_back_addr":"192.168.123.101:6809/2152256884","heartbeat_front_addr":"192.168.123.101:6807/2152256884","state":["exists","up"]},{"osd":1,"uuid":"423e15ff-ce61-4730-b91e-0f9ca81dd61d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":14,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6810","nonce":1960293858},{"type":"v1","addr":"192.168.123.101:6811","nonce":1960293858}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6812","nonce":1960293858},{"type":"v1","addr":"192.168.123.101:6813","nonce":1960293858}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6816","nonce":1960293858},{"type":"v1","addr":"192.168.123.101:6817","nonce":1960293858}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6814","nonce":1960293858},{"type":"v1","addr":"192.168.123.101:6815","nonce":1960293858}]},"public_addr":"192.168.123.101:6811/1960293858","cluster_addr":"192.168.123.101:6813/1960293858","heartbeat_back_addr":"192.168.123.101:6817/1960293858","heartbeat_front_addr":"192.168.123.101:6815/1960293858","state":["exists","up"]},{"osd":2,"uuid":"e19eb7fe-8764-496e-ad96-eed37a0f93ad","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":45,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6800","nonce":1562513062},{"type":"v1","addr":"192.168.123.102:6801","nonce":1562513062}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":1562513062},{"type":"v1","addr":"192.168.123.102:6803","nonce":1562513062}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":1562513062},{"type":"v1","addr":"192.168.123.102:6807","nonce":1562513062}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":1562513062},{"type":"v1","addr":"192.168.123.102:6805","nonce":1562513062}]},"public_addr":"192.168.123.102:6801/1562513062","cluster_addr":"192.168.123.102:6803/1562513062","heartbeat_back_addr":"192.168.123.102:6807/1562513062","heartbeat_front_addr":"192.168.123.102:6805/1562513062","state":["exists","up"]},{"osd":3,"uuid":"9916ed76-ed97-49db-960f-84cec5746a36","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":45,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":914771124},{"type":"v1","addr":"192.168.123.102:6809","nonce":914771124}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6810","nonce":914771124},{"type":"v1","addr":"192.168.123.102:6811","nonce":914771124}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6814","nonce":914771124},{"type":"v1","addr":"192.168.123.102:6815","nonce":914771124}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":914771124},{"type":"v1","addr":"192.168.123.102:6813","nonce":914771124}]},"public_addr":"192.168.123.102:6809/914771124","cluster_addr":"192.168.123.102:6811/914771124","heartbeat_back_addr":"192.168.123.102:6815/914771124","heartbeat_front_addr":"192.168.123.102:6813/914771124","state":["exists","up"]},{"osd":4,"uuid":"cb979506-99a8-4950-b248-04eeab920b11","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":30,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":4203989344},{"type":"v1","addr":"192.168.123.102:6817","nonce":4203989344}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6818","nonce":4203989344},{"type":"v1","addr":"192.168.123.102:6819","nonce":4203989344}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6822","nonce":4203989344},{"type":"v1","addr":"192.168.123.102:6823","nonce":4203989344}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6820","nonce":4203989344},{"type":"v1","addr":"192.168.123.102:6821","nonce":4203989344}]},"public_addr":"192.168.123.102:6817/4203989344","cluster_addr":"192.168.123.102:6819/4203989344","heartbeat_back_addr":"192.168.123.102:6823/4203989344","heartbeat_front_addr":"192.168.123.102:6821/4203989344","state":["exists","up"]},{"osd":5,"uuid":"4615618e-f690-4722-b0c7-d3d66c66ee61","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":34,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6800","nonce":576190648},{"type":"v1","addr":"192.168.123.105:6801","nonce":576190648}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6802","nonce":576190648},{"type":"v1","addr":"192.168.123.105:6803","nonce":576190648}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6806","nonce":576190648},{"type":"v1","addr":"192.168.123.105:6807","nonce":576190648}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6804","nonce":576190648},{"type":"v1","addr":"192.168.123.105:6805","nonce":576190648}]},"public_addr":"192.168.123.105:6801/576190648","cluster_addr":"192.168.123.105:6803/576190648","heartbeat_back_addr":"192.168.123.105:6807/576190648","heartbeat_front_addr":"192.168.123.105:6805/576190648","state":["exists","up"]},{"osd":6,"uuid":"33f0f20d-1d48-4480-9474-610fa90103f6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":39,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6808","nonce":3164767821},{"type":"v1","addr":"192.168.123.105:6809","nonce":3164767821}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6810","nonce":3164767821},{"type":"v1","addr":"192.168.123.105:6811","nonce":3164767821}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6814","nonce":3164767821},{"type":"v1","addr":"192.168.123.105:6815","nonce":3164767821}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6812","nonce":3164767821},{"type":"v1","addr":"192.168.123.105:6813","nonce":3164767821}]},"public_addr":"192.168.123.105:6809/3164767821","cluster_addr":"192.168.123.105:6811/3164767821","heartbeat_back_addr":"192.168.123.105:6815/3164767821","heartbeat_front_addr":"192.168.123.105:6813/3164767821","state":["exists","up"]},{"osd":7,"uuid":"5e5804cd-52e9-44e4-9407-776a5caf39a6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":43,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6816","nonce":2481295119},{"type":"v1","addr":"192.168.123.105:6817","nonce":2481295119}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6818","nonce":2481295119},{"type":"v1","addr":"192.168.123.105:6819","nonce":2481295119}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6822","nonce":2481295119},{"type":"v1","addr":"192.168.123.105:6823","nonce":2481295119}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.105:6820","nonce":2481295119},{"type":"v1","addr":"192.168.123.105:6821","nonce":2481295119}]},"public_addr":"192.168.123.105:6817/2481295119","cluster_addr":"192.168.123.105:6819/2481295119","heartbeat_back_addr":"192.168.123.105:6823/2481295119","heartbeat_front_addr":"192.168.123.105:6821/2481295119","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:29:51.081158+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:01.736940+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:11.825689+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:22.626076+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:33.635533+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:42.045408+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:30:52.918124+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T05:31:03.110766+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.101:6801/798997338":"2026-03-11T05:29:09.828614+0000","192.168.123.101:0/2671353675":"2026-03-11T05:29:09.828614+0000","192.168.123.101:0/1596225986":"2026-03-11T05:28:59.352838+0000","192.168.123.101:0/1037707188":"2026-03-11T05:28:59.352838+0000","192.168.123.101:6800/798997338":"2026-03-11T05:29:09.828614+0000","192.168.123.101:0/2614637726":"2026-03-11T05:28:59.352838+0000","192.168.123.101:6801/1652942659":"2026-03-11T05:28:59.352838+0000","192.168.123.101:6800/1652942659":"2026-03-11T05:28:59.352838+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T05:31:13.852 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph tell osd.0 flush_pg_stats 2026-03-10T05:31:13.853 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph tell osd.1 flush_pg_stats 2026-03-10T05:31:13.853 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph tell osd.2 flush_pg_stats 2026-03-10T05:31:13.853 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph tell osd.3 flush_pg_stats 2026-03-10T05:31:13.853 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph tell osd.4 flush_pg_stats 2026-03-10T05:31:13.853 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph tell osd.5 flush_pg_stats 2026-03-10T05:31:13.853 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph tell osd.6 flush_pg_stats 2026-03-10T05:31:13.853 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph tell osd.7 flush_pg_stats 2026-03-10T05:31:13.914 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-mon[47941]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-10T05:31:13.914 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-mon[47941]: Adding iSCSI gateway http://:@192.168.123.101:5000 to Dashboard 2026-03-10T05:31:13.914 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-mon[47941]: Adding iSCSI gateway http://:@192.168.123.105:5000 to Dashboard 2026-03-10T05:31:13.914 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-mon[47941]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-10T05:31:13.914 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-mon[47941]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm01"}]: dispatch 2026-03-10T05:31:13.914 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-mon[47941]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm05"}]: dispatch 2026-03-10T05:31:13.914 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:13.914 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:13.914 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3331582639' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:31:13.914 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3510290556' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1488281479"}]': finished 2026-03-10T05:31:13.914 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-mon[47941]: osdmap e50: 8 total, 8 up, 8 in 2026-03-10T05:31:13.914 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3150489092' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/798997338"}]: dispatch 2026-03-10T05:31:13.914 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:13 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1426613967' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:31:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:13 vm05 ceph-mon[50927]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-10T05:31:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:13 vm05 ceph-mon[50927]: Adding iSCSI gateway http://:@192.168.123.101:5000 to Dashboard 2026-03-10T05:31:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:13 vm05 ceph-mon[50927]: Adding iSCSI gateway http://:@192.168.123.105:5000 to Dashboard 2026-03-10T05:31:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:13 vm05 ceph-mon[50927]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-10T05:31:14.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:13 vm05 ceph-mon[50927]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm01"}]: dispatch 2026-03-10T05:31:14.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:13 vm05 ceph-mon[50927]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm05"}]: dispatch 2026-03-10T05:31:14.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:13 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:14.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:13 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:14.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:13 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3331582639' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:31:14.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:13 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3510290556' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1488281479"}]': finished 2026-03-10T05:31:14.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:13 vm05 ceph-mon[50927]: osdmap e50: 8 total, 8 up, 8 in 2026-03-10T05:31:14.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:13 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3150489092' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/798997338"}]: dispatch 2026-03-10T05:31:14.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:13 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1426613967' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:31:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:13 vm02 ceph-mon[50473]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-10T05:31:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:13 vm02 ceph-mon[50473]: Adding iSCSI gateway http://:@192.168.123.101:5000 to Dashboard 2026-03-10T05:31:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:13 vm02 ceph-mon[50473]: Adding iSCSI gateway http://:@192.168.123.105:5000 to Dashboard 2026-03-10T05:31:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:13 vm02 ceph-mon[50473]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-10T05:31:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:13 vm02 ceph-mon[50473]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm01"}]: dispatch 2026-03-10T05:31:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:13 vm02 ceph-mon[50473]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm05"}]: dispatch 2026-03-10T05:31:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:13 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:31:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:13 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:13 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3331582639' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:31:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:13 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3510290556' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1488281479"}]': finished 2026-03-10T05:31:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:13 vm02 ceph-mon[50473]: osdmap e50: 8 total, 8 up, 8 in 2026-03-10T05:31:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:13 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3150489092' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/798997338"}]: dispatch 2026-03-10T05:31:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:13 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1426613967' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:31:14.612 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:14.688 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:14.692 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:14.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Successfully removed blocklist entry 2026-03-10T05:31:14.700 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Removing blocklisted entry for this host : 192.168.123.101:0/2671353675 2026-03-10T05:31:14.714 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:14.811 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:14.818 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:14.824 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:15.074 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:15.215 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Successfully removed blocklist entry 2026-03-10T05:31:15.215 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Removing blocklisted entry for this host : 192.168.123.101:0/1596225986 2026-03-10T05:31:15.273 INFO:teuthology.orchestra.run.vm01.stdout:146028888072 2026-03-10T05:31:15.273 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd last-stat-seq osd.5 2026-03-10T05:31:15.361 INFO:teuthology.orchestra.run.vm01.stdout:184683593731 2026-03-10T05:31:15.361 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd last-stat-seq osd.7 2026-03-10T05:31:15.483 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:15 vm01 ceph-mon[47941]: pgmap v102: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 387 B/s wr, 0 op/s 2026-03-10T05:31:15.483 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:15 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3150489092' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/798997338"}]': finished 2026-03-10T05:31:15.483 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:15 vm01 ceph-mon[47941]: mgrmap e14: x(active, since 2m) 2026-03-10T05:31:15.483 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:15 vm01 ceph-mon[47941]: osdmap e51: 8 total, 8 up, 8 in 2026-03-10T05:31:15.484 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:15 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3563797362' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2671353675"}]: dispatch 2026-03-10T05:31:15.484 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:15 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3563797362' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2671353675"}]': finished 2026-03-10T05:31:15.484 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:15 vm01 ceph-mon[47941]: osdmap e52: 8 total, 8 up, 8 in 2026-03-10T05:31:15.484 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:15 vm02 ceph-mon[50473]: pgmap v102: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 387 B/s wr, 0 op/s 2026-03-10T05:31:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:15 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3150489092' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/798997338"}]': finished 2026-03-10T05:31:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:15 vm02 ceph-mon[50473]: mgrmap e14: x(active, since 2m) 2026-03-10T05:31:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:15 vm02 ceph-mon[50473]: osdmap e51: 8 total, 8 up, 8 in 2026-03-10T05:31:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:15 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3563797362' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2671353675"}]: dispatch 2026-03-10T05:31:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:15 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3563797362' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2671353675"}]': finished 2026-03-10T05:31:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:15 vm02 ceph-mon[50473]: osdmap e52: 8 total, 8 up, 8 in 2026-03-10T05:31:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:15.687 INFO:teuthology.orchestra.run.vm01.stdout:81604378637 2026-03-10T05:31:15.687 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd last-stat-seq osd.2 2026-03-10T05:31:15.748 INFO:teuthology.orchestra.run.vm01.stdout:107374182412 2026-03-10T05:31:15.748 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd last-stat-seq osd.3 2026-03-10T05:31:15.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:15 vm05 ceph-mon[50927]: pgmap v102: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 387 B/s wr, 0 op/s 2026-03-10T05:31:15.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:15 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3150489092' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/798997338"}]': finished 2026-03-10T05:31:15.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:15 vm05 ceph-mon[50927]: mgrmap e14: x(active, since 2m) 2026-03-10T05:31:15.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:15 vm05 ceph-mon[50927]: osdmap e51: 8 total, 8 up, 8 in 2026-03-10T05:31:15.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:15 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3563797362' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2671353675"}]: dispatch 2026-03-10T05:31:15.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:15 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3563797362' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2671353675"}]': finished 2026-03-10T05:31:15.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:15 vm05 ceph-mon[50927]: osdmap e52: 8 total, 8 up, 8 in 2026-03-10T05:31:15.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:15.806 INFO:teuthology.orchestra.run.vm01.stdout:167503724550 2026-03-10T05:31:15.807 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd last-stat-seq osd.6 2026-03-10T05:31:15.834 INFO:teuthology.orchestra.run.vm01.stdout:128849018890 2026-03-10T05:31:15.834 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd last-stat-seq osd.4 2026-03-10T05:31:15.842 INFO:teuthology.orchestra.run.vm01.stdout:38654705682 2026-03-10T05:31:15.842 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd last-stat-seq osd.0 2026-03-10T05:31:15.850 INFO:teuthology.orchestra.run.vm01.stdout:60129542160 2026-03-10T05:31:15.850 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph osd last-stat-seq osd.1 2026-03-10T05:31:16.077 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:16.149 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:16.593 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:16 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/841149437' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1596225986"}]: dispatch 2026-03-10T05:31:16.593 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:16 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Successfully removed blocklist entry 2026-03-10T05:31:16.593 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:16 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Removing blocklisted entry for this host : 192.168.123.101:0/1037707188 2026-03-10T05:31:16.609 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:16.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:16 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/841149437' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1596225986"}]: dispatch 2026-03-10T05:31:16.712 INFO:teuthology.orchestra.run.vm01.stdout:146028888072 2026-03-10T05:31:16.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:16 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/841149437' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1596225986"}]: dispatch 2026-03-10T05:31:16.852 INFO:teuthology.orchestra.run.vm01.stdout:184683593732 2026-03-10T05:31:16.891 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:16.917 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:16.957 INFO:tasks.cephadm.ceph_manager.ceph:need seq 184683593731 got 184683593732 for osd.7 2026-03-10T05:31:16.957 DEBUG:teuthology.parallel:result is None 2026-03-10T05:31:17.012 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:17.019 INFO:tasks.cephadm.ceph_manager.ceph:need seq 146028888072 got 146028888072 for osd.5 2026-03-10T05:31:17.019 DEBUG:teuthology.parallel:result is None 2026-03-10T05:31:17.045 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:17.059 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:17.105 INFO:teuthology.orchestra.run.vm01.stdout:107374182412 2026-03-10T05:31:17.247 INFO:tasks.cephadm.ceph_manager.ceph:need seq 107374182412 got 107374182412 for osd.3 2026-03-10T05:31:17.247 DEBUG:teuthology.parallel:result is None 2026-03-10T05:31:17.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:17 vm01 ceph-mon[47941]: pgmap v105: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-10T05:31:17.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:17 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/841149437' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1596225986"}]': finished 2026-03-10T05:31:17.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:17 vm01 ceph-mon[47941]: osdmap e53: 8 total, 8 up, 8 in 2026-03-10T05:31:17.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:17 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/117129574' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T05:31:17.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:17 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4090412947' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1037707188"}]: dispatch 2026-03-10T05:31:17.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:17 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/604445456' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T05:31:17.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:17 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4187445331' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T05:31:17.589 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:17 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Successfully removed blocklist entry 2026-03-10T05:31:17.589 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:17 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Removing blocklisted entry for this host : 192.168.123.101:6800/798997338 2026-03-10T05:31:17.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:17 vm02 ceph-mon[50473]: pgmap v105: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-10T05:31:17.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:17 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/841149437' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1596225986"}]': finished 2026-03-10T05:31:17.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:17 vm02 ceph-mon[50473]: osdmap e53: 8 total, 8 up, 8 in 2026-03-10T05:31:17.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:17 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/117129574' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T05:31:17.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:17 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4090412947' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1037707188"}]: dispatch 2026-03-10T05:31:17.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:17 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/604445456' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T05:31:17.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:17 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4187445331' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T05:31:17.679 INFO:teuthology.orchestra.run.vm01.stdout:81604378638 2026-03-10T05:31:17.746 INFO:teuthology.orchestra.run.vm01.stdout:167503724550 2026-03-10T05:31:17.764 INFO:tasks.cephadm.ceph_manager.ceph:need seq 81604378637 got 81604378638 for osd.2 2026-03-10T05:31:17.764 DEBUG:teuthology.parallel:result is None 2026-03-10T05:31:17.764 INFO:teuthology.orchestra.run.vm01.stdout:38654705682 2026-03-10T05:31:17.802 INFO:teuthology.orchestra.run.vm01.stdout:128849018890 2026-03-10T05:31:17.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:17 vm05 ceph-mon[50927]: pgmap v105: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-10T05:31:17.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:17 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/841149437' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1596225986"}]': finished 2026-03-10T05:31:17.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:17 vm05 ceph-mon[50927]: osdmap e53: 8 total, 8 up, 8 in 2026-03-10T05:31:17.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:17 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/117129574' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T05:31:17.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:17 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4090412947' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1037707188"}]: dispatch 2026-03-10T05:31:17.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:17 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/604445456' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T05:31:17.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:17 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4187445331' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T05:31:17.849 INFO:tasks.cephadm.ceph_manager.ceph:need seq 38654705682 got 38654705682 for osd.0 2026-03-10T05:31:17.849 DEBUG:teuthology.parallel:result is None 2026-03-10T05:31:17.872 INFO:tasks.cephadm.ceph_manager.ceph:need seq 167503724550 got 167503724550 for osd.6 2026-03-10T05:31:17.872 DEBUG:teuthology.parallel:result is None 2026-03-10T05:31:17.875 INFO:teuthology.orchestra.run.vm01.stdout:60129542160 2026-03-10T05:31:17.920 INFO:tasks.cephadm.ceph_manager.ceph:need seq 60129542160 got 60129542160 for osd.1 2026-03-10T05:31:17.920 DEBUG:teuthology.parallel:result is None 2026-03-10T05:31:17.946 INFO:tasks.cephadm.ceph_manager.ceph:need seq 128849018890 got 128849018890 for osd.4 2026-03-10T05:31:17.947 DEBUG:teuthology.parallel:result is None 2026-03-10T05:31:17.947 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-10T05:31:17.947 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph pg dump --format=json 2026-03-10T05:31:18.146 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:18.368 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:31:18.369 INFO:teuthology.orchestra.run.vm01.stderr:dumped all 2026-03-10T05:31:18.447 INFO:teuthology.orchestra.run.vm01.stdout:{"pg_ready":true,"pg_map":{"version":108,"stamp":"2026-03-10T05:31:17.855965+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459688,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":61,"num_read_kb":46,"num_write":63,"num_write_kb":587,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":41,"ondisk_log_size":41,"up":12,"acting":12,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":12,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":6,"kb":167739392,"kb_used":220116,"kb_used_data":5376,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167519276,"statfs":{"total":171765137408,"available":171539738624,"internally_reserved":0,"allocated":5505024,"data_stored":3093300,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12711,"internal_metadata":219663961},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":389,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":15,"num_read_kb":9,"num_write":4,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.000579"},"pg_stats":[{"pgid":"2.2","version":"48'2","reported_seq":29,"reported_epoch":52,"state":"active+clean","last_fresh":"2026-03-10T05:31:15.497620+0000","last_change":"2026-03-10T05:31:09.573162+0000","last_active":"2026-03-10T05:31:15.497620+0000","last_peered":"2026-03-10T05:31:15.497620+0000","last_clean":"2026-03-10T05:31:15.497620+0000","last_became_active":"2026-03-10T05:31:07.564041+0000","last_became_peered":"2026-03-10T05:31:07.564041+0000","last_unstale":"2026-03-10T05:31:15.497620+0000","last_undegraded":"2026-03-10T05:31:15.497620+0000","last_fullsized":"2026-03-10T05:31:15.497620+0000","mapping_epoch":45,"log_start":"0'0","ondisk_log_start":"0'0","created":45,"last_epoch_clean":46,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T05:31:06.542532+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T05:31:06.542532+0000","last_clean_scrub_stamp":"2026-03-10T05:31:06.542532+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:19:44.829277+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00078803499999999997,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,2],"acting":[3,1,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"2.1","version":"46'1","reported_seq":28,"reported_epoch":52,"state":"active+clean","last_fresh":"2026-03-10T05:31:14.980225+0000","last_change":"2026-03-10T05:31:09.568613+0000","last_active":"2026-03-10T05:31:14.980225+0000","last_peered":"2026-03-10T05:31:14.980225+0000","last_clean":"2026-03-10T05:31:14.980225+0000","last_became_active":"2026-03-10T05:31:07.560547+0000","last_became_peered":"2026-03-10T05:31:07.560547+0000","last_unstale":"2026-03-10T05:31:14.980225+0000","last_undegraded":"2026-03-10T05:31:14.980225+0000","last_fullsized":"2026-03-10T05:31:14.980225+0000","mapping_epoch":45,"log_start":"0'0","ondisk_log_start":"0'0","created":45,"last_epoch_clean":46,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T05:31:06.542532+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T05:31:06.542532+0000","last_clean_scrub_stamp":"2026-03-10T05:31:06.542532+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:34:26.558291+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00025752199999999999,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,1,0],"acting":[2,1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"2.0","version":"48'6","reported_seq":46,"reported_epoch":52,"state":"active+clean","last_fresh":"2026-03-10T05:31:15.497653+0000","last_change":"2026-03-10T05:31:09.573388+0000","last_active":"2026-03-10T05:31:15.497653+0000","last_peered":"2026-03-10T05:31:15.497653+0000","last_clean":"2026-03-10T05:31:15.497653+0000","last_became_active":"2026-03-10T05:31:08.155610+0000","last_became_peered":"2026-03-10T05:31:08.155610+0000","last_unstale":"2026-03-10T05:31:15.497653+0000","last_undegraded":"2026-03-10T05:31:15.497653+0000","last_fullsized":"2026-03-10T05:31:15.497653+0000","mapping_epoch":45,"log_start":"0'0","ondisk_log_start":"0'0","created":45,"last_epoch_clean":46,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T05:31:06.542532+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T05:31:06.542532+0000","last_clean_scrub_stamp":"2026-03-10T05:31:06.542532+0000","objects_scrubbed":0,"log_size":6,"log_dups_size":0,"ondisk_log_size":6,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:10:54.876034+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00094479899999999996,"stat_sum":{"num_bytes":389,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":15,"num_read_kb":9,"num_write":4,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,6],"acting":[3,5,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"1.0","version":"21'32","reported_seq":75,"reported_epoch":52,"state":"active+clean","last_fresh":"2026-03-10T05:31:15.497618+0000","last_change":"2026-03-10T05:30:46.292954+0000","last_active":"2026-03-10T05:31:15.497618+0000","last_peered":"2026-03-10T05:31:15.497618+0000","last_clean":"2026-03-10T05:31:15.497618+0000","last_became_active":"2026-03-10T05:30:45.985780+0000","last_became_peered":"2026-03-10T05:30:45.985780+0000","last_unstale":"2026-03-10T05:31:15.497618+0000","last_undegraded":"2026-03-10T05:31:15.497618+0000","last_fullsized":"2026-03-10T05:31:15.497618+0000","mapping_epoch":35,"log_start":"0'0","ondisk_log_start":"0'0","created":20,"last_epoch_clean":36,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T05:30:15.959753+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T05:30:15.959753+0000","last_clean_scrub_stamp":"2026-03-10T05:30:15.959753+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T06:11:25.440063+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,2],"acting":[3,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]}],"pool_stats":[{"poolid":2,"num_pg":3,"stat_sum":{"num_bytes":408,"num_objects":3,"num_object_clones":0,"num_object_copies":9,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":3,"num_whiteouts":0,"num_read":15,"num_read_kb":9,"num_write":6,"num_write_kb":3,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":24576,"data_stored":1224,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":9,"ondisk_log_size":9,"up":9,"acting":9,"num_store_stats":6},{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":2314240,"data_stored":2296400,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":5}],"osd_stats":[{"osd":7,"up_from":43,"seq":184683593732,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27344,"kb_used_data":504,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940080,"statfs":{"total":21470642176,"available":21442641920,"internally_reserved":0,"allocated":516096,"data_stored":216471,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1585,"internal_metadata":27457999},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":39,"seq":167503724550,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27352,"kb_used_data":508,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940072,"statfs":{"total":21470642176,"available":21442633728,"internally_reserved":0,"allocated":520192,"data_stored":216860,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":34,"seq":146028888072,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27792,"kb_used_data":948,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939632,"statfs":{"total":21470642176,"available":21442183168,"internally_reserved":0,"allocated":970752,"data_stored":670296,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":30,"seq":128849018890,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27332,"kb_used_data":492,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940092,"statfs":{"total":21470642176,"available":21442654208,"internally_reserved":0,"allocated":503808,"data_stored":210627,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":25,"seq":107374182412,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27792,"kb_used_data":952,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939632,"statfs":{"total":21470642176,"available":21442183168,"internally_reserved":0,"allocated":974848,"data_stored":670315,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":19,"seq":81604378638,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27800,"kb_used_data":960,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939624,"statfs":{"total":21470642176,"available":21442174976,"internally_reserved":0,"allocated":983040,"data_stored":675770,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":14,"seq":60129542160,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27352,"kb_used_data":508,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940072,"statfs":{"total":21470642176,"available":21442633728,"internally_reserved":0,"allocated":520192,"data_stored":216490,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":9,"seq":38654705682,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27352,"kb_used_data":504,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940072,"statfs":{"total":21470642176,"available":21442633728,"internally_reserved":0,"allocated":516096,"data_stored":216471,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":408,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-10T05:31:18.448 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph pg dump --format=json 2026-03-10T05:31:18.483 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:18 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4090412947' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1037707188"}]': finished 2026-03-10T05:31:18.483 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:18 vm01 ceph-mon[47941]: osdmap e54: 8 total, 8 up, 8 in 2026-03-10T05:31:18.483 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:18 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1526251605' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/798997338"}]: dispatch 2026-03-10T05:31:18.483 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:18 vm01 ceph-mon[47941]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/798997338"}]: dispatch 2026-03-10T05:31:18.483 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:18 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/778172110' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T05:31:18.483 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:18 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/624241813' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T05:31:18.483 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:18 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4236405965' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T05:31:18.483 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:18 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3077157028' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T05:31:18.483 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:18 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1544354269' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T05:31:18.483 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:18 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Successfully removed blocklist entry 2026-03-10T05:31:18.484 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:18 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Removing blocklisted entry for this host : 192.168.123.101:0/2614637726 2026-03-10T05:31:18.626 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:18.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:18 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4090412947' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1037707188"}]': finished 2026-03-10T05:31:18.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:18 vm05 ceph-mon[50927]: osdmap e54: 8 total, 8 up, 8 in 2026-03-10T05:31:18.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:18 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1526251605' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/798997338"}]: dispatch 2026-03-10T05:31:18.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:18 vm05 ceph-mon[50927]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/798997338"}]: dispatch 2026-03-10T05:31:18.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:18 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/778172110' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T05:31:18.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:18 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/624241813' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T05:31:18.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:18 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4236405965' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T05:31:18.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:18 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3077157028' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T05:31:18.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:18 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1544354269' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T05:31:18.847 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:31:18.848 INFO:teuthology.orchestra.run.vm01.stderr:dumped all 2026-03-10T05:31:18.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:18 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4090412947' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/1037707188"}]': finished 2026-03-10T05:31:18.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:18 vm02 ceph-mon[50473]: osdmap e54: 8 total, 8 up, 8 in 2026-03-10T05:31:18.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:18 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1526251605' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/798997338"}]: dispatch 2026-03-10T05:31:18.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:18 vm02 ceph-mon[50473]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/798997338"}]: dispatch 2026-03-10T05:31:18.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:18 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/778172110' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T05:31:18.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:18 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/624241813' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T05:31:18.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:18 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4236405965' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T05:31:18.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:18 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3077157028' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T05:31:18.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:18 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1544354269' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T05:31:18.910 INFO:teuthology.orchestra.run.vm01.stdout:{"pg_ready":true,"pg_map":{"version":109,"stamp":"2026-03-10T05:31:18.400452+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459688,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":61,"num_read_kb":46,"num_write":63,"num_write_kb":587,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":41,"ondisk_log_size":41,"up":12,"acting":12,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":12,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":6,"kb":167739392,"kb_used":220116,"kb_used_data":5376,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167519276,"statfs":{"total":171765137408,"available":171539738624,"internally_reserved":0,"allocated":5505024,"data_stored":3093300,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12711,"internal_metadata":219663961},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":389,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":15,"num_read_kb":9,"num_write":4,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.031731"},"pg_stats":[{"pgid":"2.2","version":"48'2","reported_seq":29,"reported_epoch":52,"state":"active+clean","last_fresh":"2026-03-10T05:31:15.497620+0000","last_change":"2026-03-10T05:31:09.573162+0000","last_active":"2026-03-10T05:31:15.497620+0000","last_peered":"2026-03-10T05:31:15.497620+0000","last_clean":"2026-03-10T05:31:15.497620+0000","last_became_active":"2026-03-10T05:31:07.564041+0000","last_became_peered":"2026-03-10T05:31:07.564041+0000","last_unstale":"2026-03-10T05:31:15.497620+0000","last_undegraded":"2026-03-10T05:31:15.497620+0000","last_fullsized":"2026-03-10T05:31:15.497620+0000","mapping_epoch":45,"log_start":"0'0","ondisk_log_start":"0'0","created":45,"last_epoch_clean":46,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T05:31:06.542532+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T05:31:06.542532+0000","last_clean_scrub_stamp":"2026-03-10T05:31:06.542532+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:19:44.829277+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00078803499999999997,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,2],"acting":[3,1,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"2.1","version":"46'1","reported_seq":28,"reported_epoch":52,"state":"active+clean","last_fresh":"2026-03-10T05:31:14.980225+0000","last_change":"2026-03-10T05:31:09.568613+0000","last_active":"2026-03-10T05:31:14.980225+0000","last_peered":"2026-03-10T05:31:14.980225+0000","last_clean":"2026-03-10T05:31:14.980225+0000","last_became_active":"2026-03-10T05:31:07.560547+0000","last_became_peered":"2026-03-10T05:31:07.560547+0000","last_unstale":"2026-03-10T05:31:14.980225+0000","last_undegraded":"2026-03-10T05:31:14.980225+0000","last_fullsized":"2026-03-10T05:31:14.980225+0000","mapping_epoch":45,"log_start":"0'0","ondisk_log_start":"0'0","created":45,"last_epoch_clean":46,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T05:31:06.542532+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T05:31:06.542532+0000","last_clean_scrub_stamp":"2026-03-10T05:31:06.542532+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T13:34:26.558291+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00025752199999999999,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,1,0],"acting":[2,1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"2.0","version":"48'6","reported_seq":46,"reported_epoch":52,"state":"active+clean","last_fresh":"2026-03-10T05:31:15.497653+0000","last_change":"2026-03-10T05:31:09.573388+0000","last_active":"2026-03-10T05:31:15.497653+0000","last_peered":"2026-03-10T05:31:15.497653+0000","last_clean":"2026-03-10T05:31:15.497653+0000","last_became_active":"2026-03-10T05:31:08.155610+0000","last_became_peered":"2026-03-10T05:31:08.155610+0000","last_unstale":"2026-03-10T05:31:15.497653+0000","last_undegraded":"2026-03-10T05:31:15.497653+0000","last_fullsized":"2026-03-10T05:31:15.497653+0000","mapping_epoch":45,"log_start":"0'0","ondisk_log_start":"0'0","created":45,"last_epoch_clean":46,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T05:31:06.542532+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T05:31:06.542532+0000","last_clean_scrub_stamp":"2026-03-10T05:31:06.542532+0000","objects_scrubbed":0,"log_size":6,"log_dups_size":0,"ondisk_log_size":6,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T14:10:54.876034+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00094479899999999996,"stat_sum":{"num_bytes":389,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":15,"num_read_kb":9,"num_write":4,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,6],"acting":[3,5,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"1.0","version":"21'32","reported_seq":75,"reported_epoch":52,"state":"active+clean","last_fresh":"2026-03-10T05:31:15.497618+0000","last_change":"2026-03-10T05:30:46.292954+0000","last_active":"2026-03-10T05:31:15.497618+0000","last_peered":"2026-03-10T05:31:15.497618+0000","last_clean":"2026-03-10T05:31:15.497618+0000","last_became_active":"2026-03-10T05:30:45.985780+0000","last_became_peered":"2026-03-10T05:30:45.985780+0000","last_unstale":"2026-03-10T05:31:15.497618+0000","last_undegraded":"2026-03-10T05:31:15.497618+0000","last_fullsized":"2026-03-10T05:31:15.497618+0000","mapping_epoch":35,"log_start":"0'0","ondisk_log_start":"0'0","created":20,"last_epoch_clean":36,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T05:30:15.959753+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T05:30:15.959753+0000","last_clean_scrub_stamp":"2026-03-10T05:30:15.959753+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T06:11:25.440063+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,2],"acting":[3,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]}],"pool_stats":[{"poolid":2,"num_pg":3,"stat_sum":{"num_bytes":408,"num_objects":3,"num_object_clones":0,"num_object_copies":9,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":3,"num_whiteouts":0,"num_read":15,"num_read_kb":9,"num_write":6,"num_write_kb":3,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":24576,"data_stored":1224,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":9,"ondisk_log_size":9,"up":9,"acting":9,"num_store_stats":6},{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":2314240,"data_stored":2296400,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":5}],"osd_stats":[{"osd":7,"up_from":43,"seq":184683593732,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27344,"kb_used_data":504,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940080,"statfs":{"total":21470642176,"available":21442641920,"internally_reserved":0,"allocated":516096,"data_stored":216471,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1585,"internal_metadata":27457999},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":39,"seq":167503724550,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27352,"kb_used_data":508,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940072,"statfs":{"total":21470642176,"available":21442633728,"internally_reserved":0,"allocated":520192,"data_stored":216860,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":34,"seq":146028888072,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27792,"kb_used_data":948,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939632,"statfs":{"total":21470642176,"available":21442183168,"internally_reserved":0,"allocated":970752,"data_stored":670296,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":30,"seq":128849018890,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27332,"kb_used_data":492,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940092,"statfs":{"total":21470642176,"available":21442654208,"internally_reserved":0,"allocated":503808,"data_stored":210627,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":25,"seq":107374182412,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27792,"kb_used_data":952,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939632,"statfs":{"total":21470642176,"available":21442183168,"internally_reserved":0,"allocated":974848,"data_stored":670315,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":19,"seq":81604378638,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27800,"kb_used_data":960,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939624,"statfs":{"total":21470642176,"available":21442174976,"internally_reserved":0,"allocated":983040,"data_stored":675770,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":14,"seq":60129542160,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27352,"kb_used_data":508,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940072,"statfs":{"total":21470642176,"available":21442633728,"internally_reserved":0,"allocated":520192,"data_stored":216490,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":9,"seq":38654705682,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27352,"kb_used_data":504,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940072,"statfs":{"total":21470642176,"available":21442633728,"internally_reserved":0,"allocated":516096,"data_stored":216471,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":408,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-10T05:31:18.910 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-10T05:31:18.910 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-10T05:31:18.910 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-10T05:31:18.910 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph health --format=json 2026-03-10T05:31:19.107 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:19.345 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:31:19.345 INFO:teuthology.orchestra.run.vm01.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-10T05:31:19.420 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-10T05:31:19.420 INFO:tasks.cephadm:Setup complete, yielding 2026-03-10T05:31:19.420 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T05:31:19.422 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm01.local 2026-03-10T05:31:19.422 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- bash -c 'ceph orch status' 2026-03-10T05:31:19.599 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:19.635 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:19 vm01 ceph-mon[47941]: pgmap v108: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-10T05:31:19.635 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:19 vm01 ceph-mon[47941]: from='client.24500 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:31:19.635 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:19 vm01 ceph-mon[47941]: from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/798997338"}]': finished 2026-03-10T05:31:19.635 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:19 vm01 ceph-mon[47941]: osdmap e55: 8 total, 8 up, 8 in 2026-03-10T05:31:19.635 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:19 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/677986622' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2614637726"}]: dispatch 2026-03-10T05:31:19.635 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:19 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3536342474' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T05:31:19.635 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:19 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Successfully removed blocklist entry 2026-03-10T05:31:19.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:19 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Removing blocklisted entry for this host : 192.168.123.101:6801/1652942659 2026-03-10T05:31:19.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:19 vm05 ceph-mon[50927]: pgmap v108: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-10T05:31:19.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:19 vm05 ceph-mon[50927]: from='client.24500 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:31:19.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:19 vm05 ceph-mon[50927]: from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/798997338"}]': finished 2026-03-10T05:31:19.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:19 vm05 ceph-mon[50927]: osdmap e55: 8 total, 8 up, 8 in 2026-03-10T05:31:19.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:19 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/677986622' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2614637726"}]: dispatch 2026-03-10T05:31:19.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:19 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3536342474' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T05:31:19.815 INFO:teuthology.orchestra.run.vm01.stdout:Backend: cephadm 2026-03-10T05:31:19.815 INFO:teuthology.orchestra.run.vm01.stdout:Available: Yes 2026-03-10T05:31:19.815 INFO:teuthology.orchestra.run.vm01.stdout:Paused: No 2026-03-10T05:31:19.881 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- bash -c 'ceph orch ps' 2026-03-10T05:31:19.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:19 vm02 ceph-mon[50473]: pgmap v108: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-10T05:31:19.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:19 vm02 ceph-mon[50473]: from='client.24500 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:31:19.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:19 vm02 ceph-mon[50473]: from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/798997338"}]': finished 2026-03-10T05:31:19.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:19 vm02 ceph-mon[50473]: osdmap e55: 8 total, 8 up, 8 in 2026-03-10T05:31:19.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:19 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/677986622' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2614637726"}]: dispatch 2026-03-10T05:31:19.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:19 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3536342474' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T05:31:20.036 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:iscsi.iscsi.a vm01 *:5000 running (9s) 7s ago 9s - - 3.9 654f31e6858e 87eedcfb11be 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:iscsi.iscsi.b vm05 *:5000 running (8s) 7s ago 8s - - 3.9 654f31e6858e 3d96d5824f4f 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:mgr.x vm01 *:9283,8765 running (2m) 7s ago 2m - - 19.2.3-678-ge911bdeb 654f31e6858e 031e5c93146d 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:mon.a vm01 running (2m) 7s ago 2m - 2048M 19.2.3-678-ge911bdeb 654f31e6858e 11f18971109c 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:mon.b vm02 running (114s) 49s ago 114s 38.3M 2048M 19.2.3-678-ge911bdeb 654f31e6858e 4da50b6ed5ae 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:mon.c vm05 running (115s) 7s ago 115s - 2048M 19.2.3-678-ge911bdeb 654f31e6858e 88c3ee012d9d 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:osd.0 vm01 running (92s) 7s ago 92s - 4096M 19.2.3-678-ge911bdeb 654f31e6858e 55aa508bdfbd 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:osd.1 vm01 running (81s) 7s ago 81s - 4096M 19.2.3-678-ge911bdeb 654f31e6858e f279e7241163 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:osd.2 vm02 running (71s) 49s ago 71s 61.1M 1451M 19.2.3-678-ge911bdeb 654f31e6858e cd32e7233680 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:osd.3 vm02 running (60s) 49s ago 60s 32.0M 1451M 19.2.3-678-ge911bdeb 654f31e6858e c69c0cfb0df9 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:osd.4 vm02 running (50s) 49s ago 50s 12.6M 1451M 19.2.3-678-ge911bdeb 654f31e6858e dfc2f16e1637 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:osd.5 vm05 running (40s) 7s ago 40s - 1451M 19.2.3-678-ge911bdeb 654f31e6858e 0f019ca23e77 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:osd.6 vm05 running (30s) 7s ago 30s - 1451M 19.2.3-678-ge911bdeb 654f31e6858e 240f37621334 2026-03-10T05:31:20.273 INFO:teuthology.orchestra.run.vm01.stdout:osd.7 vm05 running (19s) 7s ago 19s - 1451M 19.2.3-678-ge911bdeb 654f31e6858e cf2813e55166 2026-03-10T05:31:20.340 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- bash -c 'ceph orch ls' 2026-03-10T05:31:20.415 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:20 vm01 ceph-mon[47941]: from='client.14634 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:31:20.570 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:20.680 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:20 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/677986622' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2614637726"}]': finished 2026-03-10T05:31:20.680 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:20 vm01 ceph-mon[47941]: osdmap e56: 8 total, 8 up, 8 in 2026-03-10T05:31:20.680 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:20 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3700168781' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/1652942659"}]: dispatch 2026-03-10T05:31:20.680 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:20 vm01 ceph-mon[47941]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/1652942659"}]: dispatch 2026-03-10T05:31:20.680 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:20 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:20.680 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:20 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Successfully removed blocklist entry 2026-03-10T05:31:20.680 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:20 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Removing blocklisted entry for this host : 192.168.123.101:6800/1652942659 2026-03-10T05:31:20.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:20 vm05 ceph-mon[50927]: from='client.14634 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:31:20.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:20 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/677986622' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2614637726"}]': finished 2026-03-10T05:31:20.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:20 vm05 ceph-mon[50927]: osdmap e56: 8 total, 8 up, 8 in 2026-03-10T05:31:20.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:20 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3700168781' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/1652942659"}]: dispatch 2026-03-10T05:31:20.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:20 vm05 ceph-mon[50927]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/1652942659"}]: dispatch 2026-03-10T05:31:20.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:20 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:20.808 INFO:teuthology.orchestra.run.vm01.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-10T05:31:20.808 INFO:teuthology.orchestra.run.vm01.stdout:iscsi.datapool ?:5000 2/2 8s ago 10s vm01=iscsi.a;vm05=iscsi.b;count:2 2026-03-10T05:31:20.808 INFO:teuthology.orchestra.run.vm01.stdout:mgr 1/1 8s ago 103s vm01=x;count:1 2026-03-10T05:31:20.808 INFO:teuthology.orchestra.run.vm01.stdout:mon 3/3 49s ago 119s vm01:192.168.123.101=a;vm02:192.168.123.102=b;vm05:192.168.123.105=c;count:3 2026-03-10T05:31:20.808 INFO:teuthology.orchestra.run.vm01.stdout:osd 8 49s ago - 2026-03-10T05:31:20.874 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- bash -c 'ceph orch host ls' 2026-03-10T05:31:20.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:20 vm02 ceph-mon[50473]: from='client.14634 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T05:31:20.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:20 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/677986622' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:0/2614637726"}]': finished 2026-03-10T05:31:20.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:20 vm02 ceph-mon[50473]: osdmap e56: 8 total, 8 up, 8 in 2026-03-10T05:31:20.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:20 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3700168781' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/1652942659"}]: dispatch 2026-03-10T05:31:20.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:20 vm02 ceph-mon[50473]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/1652942659"}]: dispatch 2026-03-10T05:31:20.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:20 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:31:21.038 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:21.258 INFO:teuthology.orchestra.run.vm01.stdout:HOST ADDR LABELS STATUS 2026-03-10T05:31:21.258 INFO:teuthology.orchestra.run.vm01.stdout:vm01 192.168.123.101 2026-03-10T05:31:21.258 INFO:teuthology.orchestra.run.vm01.stdout:vm02 192.168.123.102 2026-03-10T05:31:21.258 INFO:teuthology.orchestra.run.vm01.stdout:vm05 192.168.123.105 2026-03-10T05:31:21.258 INFO:teuthology.orchestra.run.vm01.stdout:3 hosts in cluster 2026-03-10T05:31:21.321 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:31:21.321 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- bash -c 'ceph orch device ls' 2026-03-10T05:31:21.492 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:31:21.590 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-mon[47941]: from='client.14649 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:21.590 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-mon[47941]: pgmap v111: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s rd, 1 op/s 2026-03-10T05:31:21.590 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-mon[47941]: from='client.24521 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:21.590 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-mon[47941]: from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/1652942659"}]': finished 2026-03-10T05:31:21.590 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-mon[47941]: osdmap e57: 8 total, 8 up, 8 in 2026-03-10T05:31:21.590 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/983681177' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/1652942659"}]: dispatch 2026-03-10T05:31:21.590 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Successfully removed blocklist entry 2026-03-10T05:31:21.590 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: * Serving Flask app 'rbd-target-api' (lazy loading) 2026-03-10T05:31:21.590 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: * Environment: production 2026-03-10T05:31:21.590 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-10T05:31:21.590 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: Use a production WSGI server instead. 2026-03-10T05:31:21.590 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: * Debug mode: off 2026-03-10T05:31:21.590 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Reading the configuration object to update local LIO configuration 2026-03-10T05:31:21.590 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Configuration does not have an entry for this host(vm01.local) - nothing to define to LIO 2026-03-10T05:31:21.590 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug * Running on all addresses. 2026-03-10T05:31:21.590 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-10T05:31:21.590 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: * Running on all addresses. 2026-03-10T05:31:21.590 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-10T05:31:21.590 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug * Running on http://[::1]:5000/ (Press CTRL+C to quit) 2026-03-10T05:31:21.590 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: * Running on http://[::1]:5000/ (Press CTRL+C to quit) 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 80s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdb hdd DWNBRSTVMM01001 20.0G Yes 80s ago 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdc hdd DWNBRSTVMM01002 20.0G Yes 80s ago 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vdd hdd DWNBRSTVMM01003 20.0G No 80s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm01 /dev/vde hdd DWNBRSTVMM01004 20.0G No 80s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm02 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 49s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm02 /dev/vdb hdd DWNBRSTVMM02001 20.0G Yes 49s ago 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm02 /dev/vdc hdd DWNBRSTVMM02002 20.0G No 49s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm02 /dev/vdd hdd DWNBRSTVMM02003 20.0G No 49s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm02 /dev/vde hdd DWNBRSTVMM02004 20.0G No 49s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm05 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 19s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm05 /dev/vdb hdd DWNBRSTVMM05001 20.0G Yes 19s ago 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm05 /dev/vdc hdd DWNBRSTVMM05002 20.0G No 19s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm05 /dev/vdd hdd DWNBRSTVMM05003 20.0G No 19s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:31:21.712 INFO:teuthology.orchestra.run.vm01.stdout:vm05 /dev/vde hdd DWNBRSTVMM05004 20.0G No 19s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T05:31:21.772 INFO:teuthology.run_tasks:Running task install... 2026-03-10T05:31:21.774 DEBUG:teuthology.task.install:project ceph 2026-03-10T05:31:21.774 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'}, 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-10T05:31:21.774 DEBUG:teuthology.task.install:config {'extra_system_packages': {'deb': ['open-iscsi', 'multipath-tools', 'python3-xmltodict', 'python3-jmespath'], 'rpm': ['iscsi-initiator-utils', 'device-mapper-multipath', 'bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-10T05:31:21.774 INFO:teuthology.task.install:Using flavor: default 2026-03-10T05:31:21.777 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-10T05:31:21.777 INFO:teuthology.task.install:extra packages: [] 2026-03-10T05:31:21.777 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['open-iscsi', 'multipath-tools', 'python3-xmltodict', 'python3-jmespath'], 'rpm': ['iscsi-initiator-utils', 'device-mapper-multipath', 'bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-10T05:31:21.777 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:31:21.778 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['open-iscsi', 'multipath-tools', 'python3-xmltodict', 'python3-jmespath'], 'rpm': ['iscsi-initiator-utils', 'device-mapper-multipath', 'bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-10T05:31:21.778 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:31:21.778 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['open-iscsi', 'multipath-tools', 'python3-xmltodict', 'python3-jmespath'], 'rpm': ['iscsi-initiator-utils', 'device-mapper-multipath', 'bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-10T05:31:21.778 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:31:21.841 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:21 vm05 ceph-mon[50927]: from='client.14649 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:21.841 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:21 vm05 ceph-mon[50927]: pgmap v111: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s rd, 1 op/s 2026-03-10T05:31:21.841 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:21 vm05 ceph-mon[50927]: from='client.24521 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:21.841 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:21 vm05 ceph-mon[50927]: from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/1652942659"}]': finished 2026-03-10T05:31:21.841 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:21 vm05 ceph-mon[50927]: osdmap e57: 8 total, 8 up, 8 in 2026-03-10T05:31:21.841 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:21 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/983681177' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/1652942659"}]: dispatch 2026-03-10T05:31:21.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:21 vm02 ceph-mon[50473]: from='client.14649 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:21.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:21 vm02 ceph-mon[50473]: pgmap v111: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s rd, 1 op/s 2026-03-10T05:31:21.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:21 vm02 ceph-mon[50473]: from='client.24521 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:21.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:21 vm02 ceph-mon[50473]: from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6801/1652942659"}]': finished 2026-03-10T05:31:21.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:21 vm02 ceph-mon[50473]: osdmap e57: 8 total, 8 up, 8 in 2026-03-10T05:31:21.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:21 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/983681177' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/1652942659"}]: dispatch 2026-03-10T05:31:22.351 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-10T05:31:22.351 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-10T05:31:22.441 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-10T05:31:22.441 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-10T05:31:22.455 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-10T05:31:22.455 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-10T05:31:22.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:22 vm05 ceph-mon[50927]: from='client.24532 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:22.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:22 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:22.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:22 vm05 ceph-mon[50927]: from='client.14673 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:22.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:22 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/983681177' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/1652942659"}]': finished 2026-03-10T05:31:22.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:22 vm05 ceph-mon[50927]: osdmap e58: 8 total, 8 up, 8 in 2026-03-10T05:31:22.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:22 vm01 ceph-mon[47941]: from='client.24532 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:22.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:22 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:22.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:22 vm01 ceph-mon[47941]: from='client.14673 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:22.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:22 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/983681177' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/1652942659"}]': finished 2026-03-10T05:31:22.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:22 vm01 ceph-mon[47941]: osdmap e58: 8 total, 8 up, 8 in 2026-03-10T05:31:22.894 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-10T05:31:22.894 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:31:22.894 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-10T05:31:22.895 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-10T05:31:22.895 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:31:22.895 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-10T05:31:22.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:22 vm02 ceph-mon[50473]: from='client.24532 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:22.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:22 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:22.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:22 vm02 ceph-mon[50473]: from='client.14673 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:22.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:22 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/983681177' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.101:6800/1652942659"}]': finished 2026-03-10T05:31:22.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:22 vm02 ceph-mon[50473]: osdmap e58: 8 total, 8 up, 8 in 2026-03-10T05:31:22.919 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, iscsi-initiator-utils, device-mapper-multipath, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-10T05:31:22.919 DEBUG:teuthology.orchestra.run.vm02:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-10T05:31:22.921 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, iscsi-initiator-utils, device-mapper-multipath, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-10T05:31:22.921 DEBUG:teuthology.orchestra.run.vm01:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-10T05:31:22.981 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-10T05:31:22.981 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T05:31:22.981 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-10T05:31:22.984 DEBUG:teuthology.orchestra.run.vm01:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-10T05:31:22.984 DEBUG:teuthology.orchestra.run.vm02:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-10T05:31:23.010 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, iscsi-initiator-utils, device-mapper-multipath, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-10T05:31:23.010 DEBUG:teuthology.orchestra.run.vm05:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-10T05:31:23.059 DEBUG:teuthology.orchestra.run.vm02:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-10T05:31:23.060 DEBUG:teuthology.orchestra.run.vm01:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-10T05:31:23.079 DEBUG:teuthology.orchestra.run.vm05:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-10T05:31:23.123 INFO:teuthology.orchestra.run.vm02.stdout:check_obsoletes = 1 2026-03-10T05:31:23.124 DEBUG:teuthology.orchestra.run.vm02:> sudo yum clean all 2026-03-10T05:31:23.124 INFO:teuthology.orchestra.run.vm01.stdout:check_obsoletes = 1 2026-03-10T05:31:23.126 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean all 2026-03-10T05:31:23.151 DEBUG:teuthology.orchestra.run.vm05:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-10T05:31:23.215 INFO:teuthology.orchestra.run.vm05.stdout:check_obsoletes = 1 2026-03-10T05:31:23.216 DEBUG:teuthology.orchestra.run.vm05:> sudo yum clean all 2026-03-10T05:31:23.291 INFO:teuthology.orchestra.run.vm02.stdout:41 files removed 2026-03-10T05:31:23.300 INFO:teuthology.orchestra.run.vm01.stdout:41 files removed 2026-03-10T05:31:23.313 DEBUG:teuthology.orchestra.run.vm02:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd iscsi-initiator-utils device-mapper-multipath bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-10T05:31:23.325 DEBUG:teuthology.orchestra.run.vm01:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd iscsi-initiator-utils device-mapper-multipath bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-10T05:31:23.389 INFO:teuthology.orchestra.run.vm05.stdout:41 files removed 2026-03-10T05:31:23.412 DEBUG:teuthology.orchestra.run.vm05:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd iscsi-initiator-utils device-mapper-multipath bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-10T05:31:23.436 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:23 vm02 ceph-mon[50473]: from='client.14679 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:23.437 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:23 vm02 ceph-mon[50473]: pgmap v114: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s rd, 1 op/s 2026-03-10T05:31:23.437 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:23 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:23.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:23 vm05 ceph-mon[50927]: from='client.14679 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:23.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:23 vm05 ceph-mon[50927]: pgmap v114: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s rd, 1 op/s 2026-03-10T05:31:23.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:23 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:23.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:23 vm01 ceph-mon[47941]: from='client.14679 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T05:31:23.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:23 vm01 ceph-mon[47941]: pgmap v114: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s rd, 1 op/s 2026-03-10T05:31:23.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:23 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:24.642 INFO:teuthology.orchestra.run.vm01.stdout:ceph packages for x86_64 72 kB/s | 84 kB 00:01 2026-03-10T05:31:24.668 INFO:teuthology.orchestra.run.vm02.stdout:ceph packages for x86_64 70 kB/s | 84 kB 00:01 2026-03-10T05:31:24.765 INFO:teuthology.orchestra.run.vm05.stdout:ceph packages for x86_64 71 kB/s | 84 kB 00:01 2026-03-10T05:31:25.623 INFO:teuthology.orchestra.run.vm02.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-10T05:31:25.627 INFO:teuthology.orchestra.run.vm01.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-10T05:31:25.748 INFO:teuthology.orchestra.run.vm05.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-10T05:31:25.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:25 vm05 ceph-mon[50927]: pgmap v115: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 1 op/s 2026-03-10T05:31:25.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:25 vm01 ceph-mon[47941]: pgmap v115: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 1 op/s 2026-03-10T05:31:25.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:25 vm02 ceph-mon[50473]: pgmap v115: 4 pgs: 4 active+clean; 449 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 1 op/s 2026-03-10T05:31:26.558 INFO:teuthology.orchestra.run.vm02.stdout:ceph source packages 2.1 kB/s | 1.9 kB 00:00 2026-03-10T05:31:26.567 INFO:teuthology.orchestra.run.vm01.stdout:ceph source packages 2.1 kB/s | 1.9 kB 00:00 2026-03-10T05:31:26.710 INFO:teuthology.orchestra.run.vm05.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-10T05:31:27.761 INFO:teuthology.orchestra.run.vm05.stdout:CentOS Stream 9 - BaseOS 8.6 MB/s | 8.9 MB 00:01 2026-03-10T05:31:27.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:27 vm05 ceph-mon[50927]: pgmap v116: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.8 KiB/s rd, 2 op/s 2026-03-10T05:31:27.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:27 vm01 ceph-mon[47941]: pgmap v116: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.8 KiB/s rd, 2 op/s 2026-03-10T05:31:27.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:27 vm02 ceph-mon[50473]: pgmap v116: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.8 KiB/s rd, 2 op/s 2026-03-10T05:31:27.963 INFO:teuthology.orchestra.run.vm02.stdout:CentOS Stream 9 - BaseOS 6.4 MB/s | 8.9 MB 00:01 2026-03-10T05:31:29.273 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - BaseOS 3.3 MB/s | 8.9 MB 00:02 2026-03-10T05:31:29.288 INFO:teuthology.orchestra.run.vm05.stdout:CentOS Stream 9 - AppStream 29 MB/s | 27 MB 00:00 2026-03-10T05:31:29.509 INFO:teuthology.orchestra.run.vm02.stdout:CentOS Stream 9 - AppStream 29 MB/s | 27 MB 00:00 2026-03-10T05:31:29.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:29 vm05 ceph-mon[50927]: pgmap v117: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T05:31:29.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:29 vm01 ceph-mon[47941]: pgmap v117: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T05:31:29.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:29 vm02 ceph-mon[50473]: pgmap v117: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-10T05:31:31.341 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - AppStream 19 MB/s | 27 MB 00:01 2026-03-10T05:31:31.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:31 vm01 ceph-mon[47941]: pgmap v118: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.1 KiB/s rd, 2 op/s 2026-03-10T05:31:31.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:31 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:31:31.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:31 vm05 ceph-mon[50927]: pgmap v118: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.1 KiB/s rd, 2 op/s 2026-03-10T05:31:31.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:31 vm02 ceph-mon[50473]: pgmap v118: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.1 KiB/s rd, 2 op/s 2026-03-10T05:31:32.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:32 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:32.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:32 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:32.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:32 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:33.418 INFO:teuthology.orchestra.run.vm02.stdout:CentOS Stream 9 - CRB 6.7 MB/s | 8.0 MB 00:01 2026-03-10T05:31:33.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:33 vm05 ceph-mon[50927]: pgmap v119: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.9 KiB/s rd, 1 op/s 2026-03-10T05:31:33.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:33 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:33.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:33 vm01 ceph-mon[47941]: pgmap v119: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.9 KiB/s rd, 1 op/s 2026-03-10T05:31:33.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:33 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:33.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:33 vm02 ceph-mon[50473]: pgmap v119: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.9 KiB/s rd, 1 op/s 2026-03-10T05:31:33.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:33 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:34.464 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - CRB 17 MB/s | 8.0 MB 00:00 2026-03-10T05:31:34.602 INFO:teuthology.orchestra.run.vm02.stdout:CentOS Stream 9 - Extras packages 56 kB/s | 20 kB 00:00 2026-03-10T05:31:35.598 INFO:teuthology.orchestra.run.vm02.stdout:Extra Packages for Enterprise Linux 22 MB/s | 20 MB 00:00 2026-03-10T05:31:35.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:35 vm05 ceph-mon[50927]: pgmap v120: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:35.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:35 vm01 ceph-mon[47941]: pgmap v120: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:35.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:35 vm02 ceph-mon[50473]: pgmap v120: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:36.102 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - Extras packages 24 kB/s | 20 kB 00:00 2026-03-10T05:31:37.037 INFO:teuthology.orchestra.run.vm01.stdout:Extra Packages for Enterprise Linux 24 MB/s | 20 MB 00:00 2026-03-10T05:31:37.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:37 vm05 ceph-mon[50927]: pgmap v121: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:37.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:37 vm01 ceph-mon[47941]: pgmap v121: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:37.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:37 vm02 ceph-mon[50473]: pgmap v121: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:39.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:39 vm05 ceph-mon[50927]: pgmap v122: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:39.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:39 vm01 ceph-mon[47941]: pgmap v122: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:39.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:39 vm02 ceph-mon[50473]: pgmap v122: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:39.980 INFO:teuthology.orchestra.run.vm02.stdout:lab-extras 64 kB/s | 50 kB 00:00 2026-03-10T05:31:40.612 INFO:teuthology.orchestra.run.vm05.stdout:CentOS Stream 9 - CRB 942 kB/s | 8.0 MB 00:08 2026-03-10T05:31:41.465 INFO:teuthology.orchestra.run.vm01.stdout:lab-extras 63 kB/s | 50 kB 00:00 2026-03-10T05:31:41.499 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:41 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:31:41.560 INFO:teuthology.orchestra.run.vm02.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T05:31:41.560 INFO:teuthology.orchestra.run.vm02.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T05:31:41.565 INFO:teuthology.orchestra.run.vm02.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-10T05:31:41.565 INFO:teuthology.orchestra.run.vm02.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-10T05:31:41.593 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:31:41.597 INFO:teuthology.orchestra.run.vm02.stdout:======================================================================================= 2026-03-10T05:31:41.597 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-10T05:31:41.597 INFO:teuthology.orchestra.run.vm02.stdout:======================================================================================= 2026-03-10T05:31:41.597 INFO:teuthology.orchestra.run.vm02.stdout:Installing: 2026-03-10T05:31:41.597 INFO:teuthology.orchestra.run.vm02.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: device-mapper-multipath x86_64 0.8.7-45.el9 baseos 156 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: iscsi-initiator-utils x86_64 6.2.1.11-0.git4b3e853.el9 baseos 392 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout:Upgrading: 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: kpartx x86_64 0.8.7-45.el9 baseos 49 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout:Installing dependencies: 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-10T05:31:41.598 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: device-mapper-multipath-libs x86_64 0.8.7-45.el9 baseos 289 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: iscsi-initiator-utils-iscsiuio x86_64 6.2.1.11-0.git4b3e853.el9 baseos 81 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: isns-utils-libs x86_64 0.101-4.el9 baseos 101 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-10T05:31:41.599 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-10T05:31:41.600 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout:Installing weak dependencies: 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout:======================================================================================= 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout:Install 140 Packages 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout:Upgrade 3 Packages 2026-03-10T05:31:41.601 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:31:41.602 INFO:teuthology.orchestra.run.vm02.stdout:Total download size: 212 M 2026-03-10T05:31:41.602 INFO:teuthology.orchestra.run.vm02.stdout:Downloading Packages: 2026-03-10T05:31:41.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:41 vm05 ceph-mon[50927]: pgmap v123: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:41.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:41 vm01 ceph-mon[47941]: pgmap v123: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:41.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:41 vm02 ceph-mon[50473]: pgmap v123: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:42.287 INFO:teuthology.orchestra.run.vm05.stdout:CentOS Stream 9 - Extras packages 24 kB/s | 20 kB 00:00 2026-03-10T05:31:42.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:42 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:42.795 INFO:teuthology.orchestra.run.vm01.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T05:31:42.796 INFO:teuthology.orchestra.run.vm01.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T05:31:42.801 INFO:teuthology.orchestra.run.vm01.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-10T05:31:42.801 INFO:teuthology.orchestra.run.vm01.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-10T05:31:42.829 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout:======================================================================================= 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout:======================================================================================= 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout:Installing: 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: device-mapper-multipath x86_64 0.8.7-45.el9 baseos 156 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: iscsi-initiator-utils x86_64 6.2.1.11-0.git4b3e853.el9 baseos 392 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout:Upgrading: 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: kpartx x86_64 0.8.7-45.el9 baseos 49 k 2026-03-10T05:31:42.834 INFO:teuthology.orchestra.run.vm01.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout:Installing dependencies: 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: device-mapper-multipath-libs x86_64 0.8.7-45.el9 baseos 289 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: iscsi-initiator-utils-iscsiuio x86_64 6.2.1.11-0.git4b3e853.el9 baseos 81 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: isns-utils-libs x86_64 0.101-4.el9 baseos 101 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-10T05:31:42.835 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout:Installing weak dependencies: 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-10T05:31:42.836 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:31:42.837 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-10T05:31:42.837 INFO:teuthology.orchestra.run.vm01.stdout:======================================================================================= 2026-03-10T05:31:42.837 INFO:teuthology.orchestra.run.vm01.stdout:Install 140 Packages 2026-03-10T05:31:42.837 INFO:teuthology.orchestra.run.vm01.stdout:Upgrade 3 Packages 2026-03-10T05:31:42.837 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:31:42.837 INFO:teuthology.orchestra.run.vm01.stdout:Total download size: 212 M 2026-03-10T05:31:42.837 INFO:teuthology.orchestra.run.vm01.stdout:Downloading Packages: 2026-03-10T05:31:42.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:42 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:42.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:42 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:43.162 INFO:teuthology.orchestra.run.vm02.stdout:(1/143): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 14 kB/s | 6.5 kB 00:00 2026-03-10T05:31:43.302 INFO:teuthology.orchestra.run.vm05.stdout:Extra Packages for Enterprise Linux 22 MB/s | 20 MB 00:00 2026-03-10T05:31:43.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:43 vm05 ceph-mon[50927]: pgmap v124: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:43.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:43 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:43.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:43 vm01 ceph-mon[47941]: pgmap v124: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:43.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:43 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:43.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:43 vm02 ceph-mon[50473]: pgmap v124: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:43.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:43 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:43.946 INFO:teuthology.orchestra.run.vm02.stdout:(2/143): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.5 MB/s | 1.2 MB 00:00 2026-03-10T05:31:44.059 INFO:teuthology.orchestra.run.vm02.stdout:(3/143): ceph-immutable-object-cache-19.2.3-678 1.3 MB/s | 145 kB 00:00 2026-03-10T05:31:44.160 INFO:teuthology.orchestra.run.vm02.stdout:(4/143): ceph-base-19.2.3-678.ge911bdeb.el9.x86 3.8 MB/s | 5.5 MB 00:01 2026-03-10T05:31:44.196 INFO:teuthology.orchestra.run.vm01.stdout:(1/143): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 14 kB/s | 6.5 kB 00:00 2026-03-10T05:31:44.284 INFO:teuthology.orchestra.run.vm02.stdout:(5/143): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 8.7 MB/s | 1.1 MB 00:00 2026-03-10T05:31:44.307 INFO:teuthology.orchestra.run.vm02.stdout:(6/143): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 9.8 MB/s | 2.4 MB 00:00 2026-03-10T05:31:44.639 INFO:teuthology.orchestra.run.vm02.stdout:(7/143): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 13 MB/s | 4.7 MB 00:00 2026-03-10T05:31:45.213 INFO:teuthology.orchestra.run.vm01.stdout:(2/143): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.1 MB/s | 1.2 MB 00:01 2026-03-10T05:31:45.307 INFO:teuthology.orchestra.run.vm02.stdout:(8/143): ceph-common-19.2.3-678.ge911bdeb.el9.x 8.4 MB/s | 22 MB 00:02 2026-03-10T05:31:45.343 INFO:teuthology.orchestra.run.vm01.stdout:(3/143): ceph-immutable-object-cache-19.2.3-678 1.1 MB/s | 145 kB 00:00 2026-03-10T05:31:45.343 INFO:teuthology.orchestra.run.vm02.stdout:(9/143): ceph-radosgw-19.2.3-678.ge911bdeb.el9. 15 MB/s | 11 MB 00:00 2026-03-10T05:31:45.468 INFO:teuthology.orchestra.run.vm02.stdout:(10/143): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 15 MB/s | 17 MB 00:01 2026-03-10T05:31:45.469 INFO:teuthology.orchestra.run.vm02.stdout:(11/143): ceph-selinux-19.2.3-678.ge911bdeb.el9 155 kB/s | 25 kB 00:00 2026-03-10T05:31:45.595 INFO:teuthology.orchestra.run.vm02.stdout:(12/143): libcephfs-devel-19.2.3-678.ge911bdeb. 265 kB/s | 34 kB 00:00 2026-03-10T05:31:45.613 INFO:teuthology.orchestra.run.vm02.stdout:(13/143): libcephfs2-19.2.3-678.ge911bdeb.el9.x 6.8 MB/s | 1.0 MB 00:00 2026-03-10T05:31:45.708 INFO:teuthology.orchestra.run.vm02.stdout:(14/143): libcephsqlite-19.2.3-678.ge911bdeb.el 1.4 MB/s | 163 kB 00:00 2026-03-10T05:31:45.730 INFO:teuthology.orchestra.run.vm02.stdout:(15/143): librados-devel-19.2.3-678.ge911bdeb.e 1.1 MB/s | 127 kB 00:00 2026-03-10T05:31:45.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:45 vm05 ceph-mon[50927]: pgmap v125: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:45.825 INFO:teuthology.orchestra.run.vm02.stdout:(16/143): libradosstriper1-19.2.3-678.ge911bdeb 4.2 MB/s | 503 kB 00:00 2026-03-10T05:31:45.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:45 vm01 ceph-mon[47941]: pgmap v125: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:45.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:45 vm02 ceph-mon[50473]: pgmap v125: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:45.931 INFO:teuthology.orchestra.run.vm01.stdout:(4/143): ceph-base-19.2.3-678.ge911bdeb.el9.x86 2.5 MB/s | 5.5 MB 00:02 2026-03-10T05:31:45.948 INFO:teuthology.orchestra.run.vm02.stdout:(17/143): python3-ceph-argparse-19.2.3-678.ge91 368 kB/s | 45 kB 00:00 2026-03-10T05:31:46.092 INFO:teuthology.orchestra.run.vm02.stdout:(18/143): python3-ceph-common-19.2.3-678.ge911b 991 kB/s | 142 kB 00:00 2026-03-10T05:31:46.115 INFO:teuthology.orchestra.run.vm02.stdout:(19/143): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 14 MB/s | 5.4 MB 00:00 2026-03-10T05:31:46.160 INFO:teuthology.orchestra.run.vm01.stdout:(5/143): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 3.0 MB/s | 2.4 MB 00:00 2026-03-10T05:31:46.182 INFO:teuthology.orchestra.run.vm01.stdout:(6/143): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 4.3 MB/s | 1.1 MB 00:00 2026-03-10T05:31:46.205 INFO:teuthology.orchestra.run.vm02.stdout:(20/143): python3-cephfs-19.2.3-678.ge911bdeb.e 1.4 MB/s | 165 kB 00:00 2026-03-10T05:31:46.249 INFO:teuthology.orchestra.run.vm02.stdout:(21/143): python3-rados-19.2.3-678.ge911bdeb.el 2.4 MB/s | 323 kB 00:00 2026-03-10T05:31:46.335 INFO:teuthology.orchestra.run.vm02.stdout:(22/143): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.3 MB/s | 303 kB 00:00 2026-03-10T05:31:46.366 INFO:teuthology.orchestra.run.vm02.stdout:(23/143): python3-rgw-19.2.3-678.ge911bdeb.el9. 859 kB/s | 100 kB 00:00 2026-03-10T05:31:46.447 INFO:teuthology.orchestra.run.vm02.stdout:(24/143): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 760 kB/s | 85 kB 00:00 2026-03-10T05:31:46.577 INFO:teuthology.orchestra.run.vm02.stdout:(25/143): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.3 MB/s | 171 kB 00:00 2026-03-10T05:31:46.638 INFO:teuthology.orchestra.run.vm02.stdout:(26/143): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 11 MB/s | 3.1 MB 00:00 2026-03-10T05:31:46.689 INFO:teuthology.orchestra.run.vm02.stdout:(27/143): ceph-grafana-dashboards-19.2.3-678.ge 279 kB/s | 31 kB 00:00 2026-03-10T05:31:46.755 INFO:teuthology.orchestra.run.vm02.stdout:(28/143): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.3 MB/s | 150 kB 00:00 2026-03-10T05:31:47.146 INFO:teuthology.orchestra.run.vm02.stdout:(29/143): ceph-mgr-dashboard-19.2.3-678.ge911bd 8.3 MB/s | 3.8 MB 00:00 2026-03-10T05:31:47.264 INFO:teuthology.orchestra.run.vm02.stdout:(30/143): ceph-mgr-modules-core-19.2.3-678.ge91 2.1 MB/s | 253 kB 00:00 2026-03-10T05:31:47.378 INFO:teuthology.orchestra.run.vm02.stdout:(31/143): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 435 kB/s | 49 kB 00:00 2026-03-10T05:31:47.421 INFO:teuthology.orchestra.run.vm02.stdout:(32/143): ceph-mgr-diskprediction-local-19.2.3- 11 MB/s | 7.4 MB 00:00 2026-03-10T05:31:47.435 INFO:teuthology.orchestra.run.vm01.stdout:(7/143): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 3.7 MB/s | 4.7 MB 00:01 2026-03-10T05:31:47.491 INFO:teuthology.orchestra.run.vm02.stdout:(33/143): ceph-prometheus-alerts-19.2.3-678.ge9 149 kB/s | 17 kB 00:00 2026-03-10T05:31:47.539 INFO:teuthology.orchestra.run.vm02.stdout:(34/143): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.5 MB/s | 299 kB 00:00 2026-03-10T05:31:47.642 INFO:teuthology.orchestra.run.vm01.stdout:(8/143): ceph-common-19.2.3-678.ge911bdeb.el9.x 5.6 MB/s | 22 MB 00:03 2026-03-10T05:31:47.654 INFO:teuthology.orchestra.run.vm05.stdout:lab-extras 64 kB/s | 50 kB 00:00 2026-03-10T05:31:47.714 INFO:teuthology.orchestra.run.vm02.stdout:(35/143): cephadm-19.2.3-678.ge911bdeb.el9.noar 3.4 MB/s | 769 kB 00:00 2026-03-10T05:31:47.761 INFO:teuthology.orchestra.run.vm01.stdout:(9/143): ceph-selinux-19.2.3-678.ge911bdeb.el9. 211 kB/s | 25 kB 00:00 2026-03-10T05:31:47.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:47 vm05 ceph-mon[50927]: pgmap v126: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:47.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:47 vm01 ceph-mon[47941]: pgmap v126: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:47.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:47 vm02 ceph-mon[50473]: pgmap v126: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:47.969 INFO:teuthology.orchestra.run.vm02.stdout:(36/143): cryptsetup-2.8.1-3.el9.x86_64.rpm 818 kB/s | 351 kB 00:00 2026-03-10T05:31:48.183 INFO:teuthology.orchestra.run.vm02.stdout:(37/143): ceph-test-19.2.3-678.ge911bdeb.el9.x8 17 MB/s | 50 MB 00:02 2026-03-10T05:31:48.185 INFO:teuthology.orchestra.run.vm02.stdout:(38/143): device-mapper-multipath-0.8.7-45.el9. 332 kB/s | 156 kB 00:00 2026-03-10T05:31:48.187 INFO:teuthology.orchestra.run.vm02.stdout:(39/143): device-mapper-multipath-libs-0.8.7-45 1.3 MB/s | 289 kB 00:00 2026-03-10T05:31:48.311 INFO:teuthology.orchestra.run.vm02.stdout:(40/143): isns-utils-libs-0.101-4.el9.x86_64.rp 811 kB/s | 101 kB 00:00 2026-03-10T05:31:48.317 INFO:teuthology.orchestra.run.vm02.stdout:(41/143): iscsi-initiator-utils-iscsiuio-6.2.1. 612 kB/s | 81 kB 00:00 2026-03-10T05:31:48.389 INFO:teuthology.orchestra.run.vm02.stdout:(42/143): ledmon-libs-1.1.0-3.el9.x86_64.rpm 523 kB/s | 40 kB 00:00 2026-03-10T05:31:48.424 INFO:teuthology.orchestra.run.vm02.stdout:(43/143): libconfig-1.7.2-9.el9.x86_64.rpm 674 kB/s | 72 kB 00:00 2026-03-10T05:31:48.520 INFO:teuthology.orchestra.run.vm02.stdout:(44/143): iscsi-initiator-utils-6.2.1.11-0.git4 1.1 MB/s | 392 kB 00:00 2026-03-10T05:31:48.531 INFO:teuthology.orchestra.run.vm02.stdout:(45/143): libgfortran-11.5.0-14.el9.x86_64.rpm 5.5 MB/s | 794 kB 00:00 2026-03-10T05:31:48.555 INFO:teuthology.orchestra.run.vm02.stdout:(46/143): mailcap-2.1.49-5.el9.noarch.rpm 961 kB/s | 33 kB 00:00 2026-03-10T05:31:48.571 INFO:teuthology.orchestra.run.vm02.stdout:(47/143): libquadmath-11.5.0-14.el9.x86_64.rpm 1.2 MB/s | 184 kB 00:00 2026-03-10T05:31:48.610 INFO:teuthology.orchestra.run.vm02.stdout:(48/143): pciutils-3.7.0-7.el9.x86_64.rpm 1.2 MB/s | 93 kB 00:00 2026-03-10T05:31:48.661 INFO:teuthology.orchestra.run.vm02.stdout:(49/143): python3-ply-3.11-14.el9.noarch.rpm 2.1 MB/s | 106 kB 00:00 2026-03-10T05:31:48.675 INFO:teuthology.orchestra.run.vm02.stdout:(50/143): python3-cffi-1.14.5-5.el9.x86_64.rpm 2.1 MB/s | 253 kB 00:00 2026-03-10T05:31:48.780 INFO:teuthology.orchestra.run.vm01.stdout:(10/143): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 6.6 MB/s | 17 MB 00:02 2026-03-10T05:31:48.801 INFO:teuthology.orchestra.run.vm02.stdout:(51/143): python3-pycparser-2.20-6.el9.noarch.r 969 kB/s | 135 kB 00:00 2026-03-10T05:31:48.816 INFO:teuthology.orchestra.run.vm02.stdout:(52/143): python3-pyparsing-2.4.7-9.el9.noarch. 1.0 MB/s | 150 kB 00:00 2026-03-10T05:31:48.867 INFO:teuthology.orchestra.run.vm02.stdout:(53/143): python3-requests-2.25.1-10.el9.noarch 1.9 MB/s | 126 kB 00:00 2026-03-10T05:31:48.900 INFO:teuthology.orchestra.run.vm01.stdout:(11/143): libcephfs-devel-19.2.3-678.ge911bdeb. 280 kB/s | 34 kB 00:00 2026-03-10T05:31:48.900 INFO:teuthology.orchestra.run.vm02.stdout:(54/143): python3-urllib3-1.26.5-7.el9.noarch.r 2.5 MB/s | 218 kB 00:00 2026-03-10T05:31:48.937 INFO:teuthology.orchestra.run.vm05.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T05:31:48.937 INFO:teuthology.orchestra.run.vm05.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T05:31:48.942 INFO:teuthology.orchestra.run.vm05.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-10T05:31:48.943 INFO:teuthology.orchestra.run.vm05.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-10T05:31:48.969 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout:======================================================================================= 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout:======================================================================================= 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout:Installing: 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: device-mapper-multipath x86_64 0.8.7-45.el9 baseos 156 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: iscsi-initiator-utils x86_64 6.2.1.11-0.git4b3e853.el9 baseos 392 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout:Upgrading: 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: kpartx x86_64 0.8.7-45.el9 baseos 49 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout:Installing dependencies: 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-10T05:31:48.974 INFO:teuthology.orchestra.run.vm05.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: device-mapper-multipath-libs x86_64 0.8.7-45.el9 baseos 289 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: iscsi-initiator-utils-iscsiuio x86_64 6.2.1.11-0.git4b3e853.el9 baseos 81 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: isns-utils-libs x86_64 0.101-4.el9 baseos 101 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-10T05:31:48.975 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout:Installing weak dependencies: 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout:======================================================================================= 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout:Install 140 Packages 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout:Upgrade 3 Packages 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout:Total download size: 212 M 2026-03-10T05:31:48.976 INFO:teuthology.orchestra.run.vm05.stdout:Downloading Packages: 2026-03-10T05:31:49.192 INFO:teuthology.orchestra.run.vm01.stdout:(12/143): libcephfs2-19.2.3-678.ge911bdeb.el9.x 3.3 MB/s | 1.0 MB 00:00 2026-03-10T05:31:49.314 INFO:teuthology.orchestra.run.vm01.stdout:(13/143): libcephsqlite-19.2.3-678.ge911bdeb.el 1.3 MB/s | 163 kB 00:00 2026-03-10T05:31:49.460 INFO:teuthology.orchestra.run.vm01.stdout:(14/143): librados-devel-19.2.3-678.ge911bdeb.e 868 kB/s | 127 kB 00:00 2026-03-10T05:31:49.466 INFO:teuthology.orchestra.run.vm02.stdout:(55/143): python3-cryptography-36.0.1-5.el9.x86 1.4 MB/s | 1.2 MB 00:00 2026-03-10T05:31:49.568 INFO:teuthology.orchestra.run.vm02.stdout:(56/143): zip-3.0-35.el9.x86_64.rpm 398 kB/s | 266 kB 00:00 2026-03-10T05:31:49.570 INFO:teuthology.orchestra.run.vm02.stdout:(57/143): unzip-6.0-59.el9.x86_64.rpm 259 kB/s | 182 kB 00:00 2026-03-10T05:31:49.585 INFO:teuthology.orchestra.run.vm01.stdout:(15/143): libradosstriper1-19.2.3-678.ge911bdeb 4.0 MB/s | 503 kB 00:00 2026-03-10T05:31:49.635 INFO:teuthology.orchestra.run.vm02.stdout:(58/143): boost-program-options-1.75.0-13.el9.x 614 kB/s | 104 kB 00:00 2026-03-10T05:31:49.653 INFO:teuthology.orchestra.run.vm02.stdout:(59/143): flexiblas-3.0.4-9.el9.x86_64.rpm 349 kB/s | 30 kB 00:00 2026-03-10T05:31:49.711 INFO:teuthology.orchestra.run.vm02.stdout:(60/143): flexiblas-openblas-openmp-3.0.4-9.el9 197 kB/s | 15 kB 00:00 2026-03-10T05:31:49.732 INFO:teuthology.orchestra.run.vm02.stdout:(61/143): libnbd-1.20.3-4.el9.x86_64.rpm 2.0 MB/s | 164 kB 00:00 2026-03-10T05:31:49.768 INFO:teuthology.orchestra.run.vm02.stdout:(62/143): libpmemobj-1.12.1-1.el9.x86_64.rpm 2.7 MB/s | 160 kB 00:00 2026-03-10T05:31:49.780 INFO:teuthology.orchestra.run.vm02.stdout:(63/143): flexiblas-netlib-3.0.4-9.el9.x86_64.r 14 MB/s | 3.0 MB 00:00 2026-03-10T05:31:49.781 INFO:teuthology.orchestra.run.vm02.stdout:(64/143): librabbitmq-0.11.0-7.el9.x86_64.rpm 925 kB/s | 45 kB 00:00 2026-03-10T05:31:49.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:49 vm05 ceph-mon[50927]: pgmap v127: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:49.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:49 vm01 ceph-mon[47941]: pgmap v127: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:49.858 INFO:teuthology.orchestra.run.vm02.stdout:(65/143): librdkafka-1.6.1-102.el9.x86_64.rpm 7.2 MB/s | 662 kB 00:00 2026-03-10T05:31:49.859 INFO:teuthology.orchestra.run.vm02.stdout:(66/143): libxslt-1.1.34-12.el9.x86_64.rpm 2.9 MB/s | 233 kB 00:00 2026-03-10T05:31:49.877 INFO:teuthology.orchestra.run.vm02.stdout:(67/143): libstoragemgmt-1.10.1-1.el9.x86_64.rp 2.5 MB/s | 246 kB 00:00 2026-03-10T05:31:49.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:49 vm02 ceph-mon[50473]: pgmap v127: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:49.937 INFO:teuthology.orchestra.run.vm02.stdout:(68/143): lua-5.4.4-4.el9.x86_64.rpm 2.4 MB/s | 188 kB 00:00 2026-03-10T05:31:49.938 INFO:teuthology.orchestra.run.vm02.stdout:(69/143): lttng-ust-2.12.0-6.el9.x86_64.rpm 3.6 MB/s | 292 kB 00:00 2026-03-10T05:31:49.939 INFO:teuthology.orchestra.run.vm02.stdout:(70/143): openblas-0.3.29-1.el9.x86_64.rpm 686 kB/s | 42 kB 00:00 2026-03-10T05:31:49.996 INFO:teuthology.orchestra.run.vm02.stdout:(71/143): protobuf-3.14.0-17.el9.x86_64.rpm 18 MB/s | 1.0 MB 00:00 2026-03-10T05:31:50.063 INFO:teuthology.orchestra.run.vm02.stdout:(72/143): python3-devel-3.9.25-3.el9.x86_64.rpm 3.6 MB/s | 244 kB 00:00 2026-03-10T05:31:50.129 INFO:teuthology.orchestra.run.vm02.stdout:(73/143): python3-jinja2-2.11.3-8.el9.noarch.rp 3.7 MB/s | 249 kB 00:00 2026-03-10T05:31:50.165 INFO:teuthology.orchestra.run.vm02.stdout:(74/143): openblas-openmp-0.3.29-1.el9.x86_64.r 23 MB/s | 5.3 MB 00:00 2026-03-10T05:31:50.183 INFO:teuthology.orchestra.run.vm02.stdout:(75/143): python3-babel-2.9.1-2.el9.noarch.rpm 24 MB/s | 6.0 MB 00:00 2026-03-10T05:31:50.184 INFO:teuthology.orchestra.run.vm02.stdout:(76/143): python3-jmespath-1.0.1-1.el9.noarch.r 863 kB/s | 48 kB 00:00 2026-03-10T05:31:50.210 INFO:teuthology.orchestra.run.vm01.stdout:(16/143): ceph-radosgw-19.2.3-678.ge911bdeb.el9 3.9 MB/s | 11 MB 00:02 2026-03-10T05:31:50.215 INFO:teuthology.orchestra.run.vm02.stdout:(77/143): python3-libstoragemgmt-1.10.1-1.el9.x 3.5 MB/s | 177 kB 00:00 2026-03-10T05:31:50.231 INFO:teuthology.orchestra.run.vm02.stdout:(78/143): python3-mako-1.1.4-6.el9.noarch.rpm 3.5 MB/s | 172 kB 00:00 2026-03-10T05:31:50.232 INFO:teuthology.orchestra.run.vm02.stdout:(79/143): python3-markupsafe-1.1.1-12.el9.x86_6 721 kB/s | 35 kB 00:00 2026-03-10T05:31:50.289 INFO:teuthology.orchestra.run.vm02.stdout:(80/143): python3-packaging-20.9-5.el9.noarch.r 1.3 MB/s | 77 kB 00:00 2026-03-10T05:31:50.296 INFO:teuthology.orchestra.run.vm02.stdout:(81/143): python3-numpy-f2py-1.23.5-2.el9.x86_6 6.7 MB/s | 442 kB 00:00 2026-03-10T05:31:50.304 INFO:teuthology.orchestra.run.vm05.stdout:(1/143): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 14 kB/s | 6.5 kB 00:00 2026-03-10T05:31:50.374 INFO:teuthology.orchestra.run.vm02.stdout:(82/143): python3-numpy-1.23.5-2.el9.x86_64.rpm 39 MB/s | 6.1 MB 00:00 2026-03-10T05:31:50.375 INFO:teuthology.orchestra.run.vm02.stdout:(83/143): python3-protobuf-3.14.0-17.el9.noarch 3.1 MB/s | 267 kB 00:00 2026-03-10T05:31:50.375 INFO:teuthology.orchestra.run.vm02.stdout:(84/143): python3-pyasn1-0.4.8-7.el9.noarch.rpm 1.9 MB/s | 157 kB 00:00 2026-03-10T05:31:50.447 INFO:teuthology.orchestra.run.vm02.stdout:(85/143): python3-pyasn1-modules-0.4.8-7.el9.no 3.8 MB/s | 277 kB 00:00 2026-03-10T05:31:50.471 INFO:teuthology.orchestra.run.vm02.stdout:(86/143): python3-requests-oauthlib-1.3.0-12.el 566 kB/s | 54 kB 00:00 2026-03-10T05:31:50.499 INFO:teuthology.orchestra.run.vm02.stdout:(87/143): python3-toml-0.10.2-6.el9.noarch.rpm 815 kB/s | 42 kB 00:00 2026-03-10T05:31:50.500 INFO:teuthology.orchestra.run.vm01.stdout:(17/143): ceph-test-19.2.3-678.ge911bdeb.el9.x8 18 MB/s | 50 MB 00:02 2026-03-10T05:31:50.533 INFO:teuthology.orchestra.run.vm02.stdout:(88/143): qatlib-25.08.0-2.el9.x86_64.rpm 3.8 MB/s | 240 kB 00:00 2026-03-10T05:31:50.583 INFO:teuthology.orchestra.run.vm02.stdout:(89/143): qatlib-service-25.08.0-2.el9.x86_64.r 445 kB/s | 37 kB 00:00 2026-03-10T05:31:50.599 INFO:teuthology.orchestra.run.vm02.stdout:(90/143): qatzip-libs-1.3.1-1.el9.x86_64.rpm 1.0 MB/s | 66 kB 00:00 2026-03-10T05:31:50.621 INFO:teuthology.orchestra.run.vm01.stdout:(18/143): python3-ceph-common-19.2.3-678.ge911b 1.1 MB/s | 142 kB 00:00 2026-03-10T05:31:50.648 INFO:teuthology.orchestra.run.vm02.stdout:(91/143): socat-1.7.4.1-8.el9.x86_64.rpm 4.6 MB/s | 303 kB 00:00 2026-03-10T05:31:50.717 INFO:teuthology.orchestra.run.vm02.stdout:(92/143): python3-scipy-1.9.3-2.el9.x86_64.rpm 57 MB/s | 19 MB 00:00 2026-03-10T05:31:50.718 INFO:teuthology.orchestra.run.vm02.stdout:(93/143): xmlstarlet-1.6.1-20.el9.x86_64.rpm 538 kB/s | 64 kB 00:00 2026-03-10T05:31:50.742 INFO:teuthology.orchestra.run.vm01.stdout:(19/143): python3-cephfs-19.2.3-678.ge911bdeb.e 1.3 MB/s | 165 kB 00:00 2026-03-10T05:31:50.772 INFO:teuthology.orchestra.run.vm01.stdout:(20/143): python3-ceph-argparse-19.2.3-678.ge91 80 kB/s | 45 kB 00:00 2026-03-10T05:31:50.789 INFO:teuthology.orchestra.run.vm02.stdout:(94/143): lua-devel-5.4.4-4.el9.x86_64.rpm 158 kB/s | 22 kB 00:00 2026-03-10T05:31:50.864 INFO:teuthology.orchestra.run.vm01.stdout:(21/143): python3-rados-19.2.3-678.ge911bdeb.el 2.6 MB/s | 323 kB 00:00 2026-03-10T05:31:50.943 INFO:teuthology.orchestra.run.vm02.stdout:(95/143): protobuf-compiler-3.14.0-17.el9.x86_6 3.7 MB/s | 862 kB 00:00 2026-03-10T05:31:50.984 INFO:teuthology.orchestra.run.vm01.stdout:(22/143): python3-rgw-19.2.3-678.ge911bdeb.el9. 833 kB/s | 100 kB 00:00 2026-03-10T05:31:50.998 INFO:teuthology.orchestra.run.vm01.stdout:(23/143): python3-rbd-19.2.3-678.ge911bdeb.el9. 1.3 MB/s | 303 kB 00:00 2026-03-10T05:31:50.999 INFO:teuthology.orchestra.run.vm02.stdout:(96/143): gperftools-libs-2.9.1-3.el9.x86_64.rp 1.4 MB/s | 308 kB 00:00 2026-03-10T05:31:51.042 INFO:teuthology.orchestra.run.vm02.stdout:(97/143): abseil-cpp-20211102.0-4.el9.x86_64.rp 1.7 MB/s | 551 kB 00:00 2026-03-10T05:31:51.079 INFO:teuthology.orchestra.run.vm02.stdout:(98/143): grpc-data-1.46.7-10.el9.noarch.rpm 144 kB/s | 19 kB 00:00 2026-03-10T05:31:51.099 INFO:teuthology.orchestra.run.vm02.stdout:(99/143): libarrow-doc-9.0.0-15.el9.noarch.rpm 436 kB/s | 25 kB 00:00 2026-03-10T05:31:51.114 INFO:teuthology.orchestra.run.vm05.stdout:(2/143): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.4 MB/s | 1.2 MB 00:00 2026-03-10T05:31:51.114 INFO:teuthology.orchestra.run.vm01.stdout:(24/143): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 656 kB/s | 85 kB 00:00 2026-03-10T05:31:51.157 INFO:teuthology.orchestra.run.vm02.stdout:(100/143): libunwind-1.6.2-1.el9.x86_64.rpm 1.2 MB/s | 67 kB 00:00 2026-03-10T05:31:51.168 INFO:teuthology.orchestra.run.vm02.stdout:(101/143): liboath-2.6.12-1.el9.x86_64.rpm 547 kB/s | 49 kB 00:00 2026-03-10T05:31:51.217 INFO:teuthology.orchestra.run.vm02.stdout:(102/143): luarocks-3.9.2-5.el9.noarch.rpm 2.5 MB/s | 151 kB 00:00 2026-03-10T05:31:51.237 INFO:teuthology.orchestra.run.vm01.stdout:(25/143): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.4 MB/s | 171 kB 00:00 2026-03-10T05:31:51.289 INFO:teuthology.orchestra.run.vm02.stdout:(103/143): python3-asyncssh-2.13.2-5.el9.noarch 7.4 MB/s | 548 kB 00:00 2026-03-10T05:31:51.339 INFO:teuthology.orchestra.run.vm02.stdout:(104/143): libarrow-9.0.0-15.el9.x86_64.rpm 13 MB/s | 4.4 MB 00:00 2026-03-10T05:31:51.352 INFO:teuthology.orchestra.run.vm02.stdout:(105/143): python3-autocommand-2.2.2-8.el9.noar 473 kB/s | 29 kB 00:00 2026-03-10T05:31:51.356 INFO:teuthology.orchestra.run.vm01.stdout:(26/143): ceph-grafana-dashboards-19.2.3-678.ge 262 kB/s | 31 kB 00:00 2026-03-10T05:31:51.407 INFO:teuthology.orchestra.run.vm02.stdout:(106/143): python3-bcrypt-3.2.2-1.el9.x86_64.rp 783 kB/s | 43 kB 00:00 2026-03-10T05:31:51.414 INFO:teuthology.orchestra.run.vm02.stdout:(107/143): python3-backports-tarfile-1.2.0-1.el 810 kB/s | 60 kB 00:00 2026-03-10T05:31:51.473 INFO:teuthology.orchestra.run.vm02.stdout:(108/143): python3-certifi-2023.05.07-4.el9.noa 240 kB/s | 14 kB 00:00 2026-03-10T05:31:51.476 INFO:teuthology.orchestra.run.vm02.stdout:(109/143): python3-cachetools-4.2.4-1.el9.noarc 467 kB/s | 32 kB 00:00 2026-03-10T05:31:51.477 INFO:teuthology.orchestra.run.vm01.stdout:(27/143): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.2 MB/s | 150 kB 00:00 2026-03-10T05:31:51.527 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:31:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:31:51.529 INFO:teuthology.orchestra.run.vm02.stdout:(110/143): python3-cheroot-10.0.1-4.el9.noarch. 3.0 MB/s | 173 kB 00:00 2026-03-10T05:31:51.540 INFO:teuthology.orchestra.run.vm02.stdout:(111/143): python3-cherrypy-18.6.1-2.el9.noarch 5.5 MB/s | 358 kB 00:00 2026-03-10T05:31:51.583 INFO:teuthology.orchestra.run.vm02.stdout:(112/143): python3-google-auth-2.45.0-1.el9.noa 4.6 MB/s | 254 kB 00:00 2026-03-10T05:31:51.638 INFO:teuthology.orchestra.run.vm01.stdout:(28/143): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 2.6 MB/s | 5.4 MB 00:02 2026-03-10T05:31:51.644 INFO:teuthology.orchestra.run.vm02.stdout:(113/143): python3-grpcio-tools-1.46.7-10.el9.x 2.3 MB/s | 144 kB 00:00 2026-03-10T05:31:51.670 INFO:teuthology.orchestra.run.vm02.stdout:(114/143): parquet-libs-9.0.0-15.el9.x86_64.rpm 1.6 MB/s | 838 kB 00:00 2026-03-10T05:31:51.702 INFO:teuthology.orchestra.run.vm02.stdout:(115/143): python3-jaraco-8.2.1-3.el9.noarch.rp 185 kB/s | 11 kB 00:00 2026-03-10T05:31:51.718 INFO:teuthology.orchestra.run.vm02.stdout:(116/143): python3-grpcio-1.46.7-10.el9.x86_64. 11 MB/s | 2.0 MB 00:00 2026-03-10T05:31:51.727 INFO:teuthology.orchestra.run.vm02.stdout:(117/143): python3-jaraco-classes-3.2.1-5.el9.n 315 kB/s | 18 kB 00:00 2026-03-10T05:31:51.740 INFO:teuthology.orchestra.run.vm01.stdout:(29/143): ceph-mgr-dashboard-19.2.3-678.ge911bd 14 MB/s | 3.8 MB 00:00 2026-03-10T05:31:51.759 INFO:teuthology.orchestra.run.vm02.stdout:(118/143): python3-jaraco-collections-3.0.0-8.e 407 kB/s | 23 kB 00:00 2026-03-10T05:31:51.773 INFO:teuthology.orchestra.run.vm02.stdout:(119/143): python3-jaraco-context-6.0.1-3.el9.n 363 kB/s | 20 kB 00:00 2026-03-10T05:31:51.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:51 vm05 ceph-mon[50927]: pgmap v128: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:51.814 INFO:teuthology.orchestra.run.vm02.stdout:(120/143): python3-jaraco-functools-3.5.0-2.el9 223 kB/s | 19 kB 00:00 2026-03-10T05:31:51.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:51 vm01 ceph-mon[47941]: pgmap v128: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:51.861 INFO:teuthology.orchestra.run.vm02.stdout:(121/143): python3-jaraco-text-4.0.0-2.el9.noar 259 kB/s | 26 kB 00:00 2026-03-10T05:31:51.865 INFO:teuthology.orchestra.run.vm01.stdout:(30/143): ceph-mgr-modules-core-19.2.3-678.ge91 2.0 MB/s | 253 kB 00:00 2026-03-10T05:31:51.888 INFO:teuthology.orchestra.run.vm02.stdout:(122/143): python3-logutils-0.3.5-21.el9.noarch 627 kB/s | 46 kB 00:00 2026-03-10T05:31:51.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:51 vm02 ceph-mon[50473]: pgmap v128: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:51.913 INFO:teuthology.orchestra.run.vm02.stdout:(123/143): python3-more-itertools-8.12.0-2.el9. 1.5 MB/s | 79 kB 00:00 2026-03-10T05:31:51.957 INFO:teuthology.orchestra.run.vm02.stdout:(124/143): python3-natsort-7.1.1-5.el9.noarch.r 841 kB/s | 58 kB 00:00 2026-03-10T05:31:51.985 INFO:teuthology.orchestra.run.vm01.stdout:(31/143): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 412 kB/s | 49 kB 00:00 2026-03-10T05:31:51.986 INFO:teuthology.orchestra.run.vm02.stdout:(125/143): python3-pecan-1.4.2-3.el9.noarch.rpm 3.6 MB/s | 272 kB 00:00 2026-03-10T05:31:52.008 INFO:teuthology.orchestra.run.vm02.stdout:(126/143): python3-portend-3.1.0-2.el9.noarch.r 321 kB/s | 16 kB 00:00 2026-03-10T05:31:52.045 INFO:teuthology.orchestra.run.vm02.stdout:(127/143): python3-pyOpenSSL-21.0.0-1.el9.noarc 1.5 MB/s | 90 kB 00:00 2026-03-10T05:31:52.060 INFO:teuthology.orchestra.run.vm02.stdout:(128/143): python3-repoze-lru-0.7-16.el9.noarch 599 kB/s | 31 kB 00:00 2026-03-10T05:31:52.102 INFO:teuthology.orchestra.run.vm02.stdout:(129/143): python3-routes-2.5.1-5.el9.noarch.rp 3.3 MB/s | 188 kB 00:00 2026-03-10T05:31:52.104 INFO:teuthology.orchestra.run.vm01.stdout:(32/143): ceph-prometheus-alerts-19.2.3-678.ge9 140 kB/s | 17 kB 00:00 2026-03-10T05:31:52.132 INFO:teuthology.orchestra.run.vm02.stdout:(130/143): python3-rsa-4.9-2.el9.noarch.rpm 819 kB/s | 59 kB 00:00 2026-03-10T05:31:52.160 INFO:teuthology.orchestra.run.vm02.stdout:(131/143): python3-kubernetes-26.1.0-3.el9.noar 2.6 MB/s | 1.0 MB 00:00 2026-03-10T05:31:52.169 INFO:teuthology.orchestra.run.vm02.stdout:(132/143): python3-tempora-5.0.0-2.el9.noarch.r 537 kB/s | 36 kB 00:00 2026-03-10T05:31:52.216 INFO:teuthology.orchestra.run.vm02.stdout:(133/143): python3-webob-1.8.8-2.el9.noarch.rpm 4.1 MB/s | 230 kB 00:00 2026-03-10T05:31:52.226 INFO:teuthology.orchestra.run.vm01.stdout:(33/143): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.4 MB/s | 299 kB 00:00 2026-03-10T05:31:52.236 INFO:teuthology.orchestra.run.vm02.stdout:(134/143): python3-typing-extensions-4.15.0-1.e 838 kB/s | 86 kB 00:00 2026-03-10T05:31:52.259 INFO:teuthology.orchestra.run.vm02.stdout:(135/143): python3-websocket-client-1.2.3-2.el9 999 kB/s | 90 kB 00:00 2026-03-10T05:31:52.278 INFO:teuthology.orchestra.run.vm02.stdout:(136/143): python3-werkzeug-2.0.3-3.el9.1.noarc 6.8 MB/s | 427 kB 00:00 2026-03-10T05:31:52.290 INFO:teuthology.orchestra.run.vm05.stdout:(3/143): ceph-immutable-object-cache-19.2.3-678 123 kB/s | 145 kB 00:01 2026-03-10T05:31:52.290 INFO:teuthology.orchestra.run.vm02.stdout:(137/143): python3-xmltodict-0.12.0-15.el9.noar 410 kB/s | 22 kB 00:00 2026-03-10T05:31:52.311 INFO:teuthology.orchestra.run.vm02.stdout:(138/143): python3-zc-lockfile-2.0-10.el9.noarc 382 kB/s | 20 kB 00:00 2026-03-10T05:31:52.336 INFO:teuthology.orchestra.run.vm02.stdout:(139/143): re2-20211101-20.el9.x86_64.rpm 3.2 MB/s | 191 kB 00:00 2026-03-10T05:31:52.355 INFO:teuthology.orchestra.run.vm01.stdout:(34/143): cephadm-19.2.3-678.ge911bdeb.el9.noar 5.9 MB/s | 769 kB 00:00 2026-03-10T05:31:52.391 INFO:teuthology.orchestra.run.vm05.stdout:(4/143): ceph-common-19.2.3-678.ge911bdeb.el9.x 8.5 MB/s | 22 MB 00:02 2026-03-10T05:31:52.516 INFO:teuthology.orchestra.run.vm05.stdout:(5/143): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 8.6 MB/s | 1.1 MB 00:00 2026-03-10T05:31:52.535 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:52 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:52.536 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:31:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:31:52.590 INFO:teuthology.orchestra.run.vm01.stdout:(35/143): cryptsetup-2.8.1-3.el9.x86_64.rpm 1.5 MB/s | 351 kB 00:00 2026-03-10T05:31:52.666 INFO:teuthology.orchestra.run.vm05.stdout:(6/143): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 6.4 MB/s | 2.4 MB 00:00 2026-03-10T05:31:52.673 INFO:teuthology.orchestra.run.vm01.stdout:(36/143): device-mapper-multipath-0.8.7-45.el9. 1.8 MB/s | 156 kB 00:00 2026-03-10T05:31:52.779 INFO:teuthology.orchestra.run.vm01.stdout:(37/143): device-mapper-multipath-libs-0.8.7-45 2.7 MB/s | 289 kB 00:00 2026-03-10T05:31:52.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:52 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:52.866 INFO:teuthology.orchestra.run.vm01.stdout:(38/143): iscsi-initiator-utils-6.2.1.11-0.git4 4.5 MB/s | 392 kB 00:00 2026-03-10T05:31:52.888 INFO:teuthology.orchestra.run.vm05.stdout:(7/143): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 13 MB/s | 4.7 MB 00:00 2026-03-10T05:31:52.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:52 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:52.930 INFO:teuthology.orchestra.run.vm01.stdout:(39/143): iscsi-initiator-utils-iscsiuio-6.2.1. 1.2 MB/s | 81 kB 00:00 2026-03-10T05:31:53.009 INFO:teuthology.orchestra.run.vm01.stdout:(40/143): isns-utils-libs-0.101-4.el9.x86_64.rp 1.2 MB/s | 101 kB 00:00 2026-03-10T05:31:53.058 INFO:teuthology.orchestra.run.vm02.stdout:(140/143): thrift-0.15.0-4.el9.x86_64.rpm 2.1 MB/s | 1.6 MB 00:00 2026-03-10T05:31:53.101 INFO:teuthology.orchestra.run.vm01.stdout:(41/143): ledmon-libs-1.1.0-3.el9.x86_64.rpm 444 kB/s | 40 kB 00:00 2026-03-10T05:31:53.155 INFO:teuthology.orchestra.run.vm01.stdout:(42/143): libconfig-1.7.2-9.el9.x86_64.rpm 1.3 MB/s | 72 kB 00:00 2026-03-10T05:31:53.191 INFO:teuthology.orchestra.run.vm02.stdout:(141/143): kpartx-0.8.7-45.el9.x86_64.rpm 371 kB/s | 49 kB 00:00 2026-03-10T05:31:53.252 INFO:teuthology.orchestra.run.vm02.stdout:(142/143): librbd1-19.2.3-678.ge911bdeb.el9.x86 3.5 MB/s | 3.2 MB 00:00 2026-03-10T05:31:53.300 INFO:teuthology.orchestra.run.vm01.stdout:(43/143): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 1.4 MB/s | 3.1 MB 00:02 2026-03-10T05:31:53.304 INFO:teuthology.orchestra.run.vm01.stdout:(44/143): libgfortran-11.5.0-14.el9.x86_64.rpm 5.2 MB/s | 794 kB 00:00 2026-03-10T05:31:53.335 INFO:teuthology.orchestra.run.vm02.stdout:(143/143): librados2-19.2.3-678.ge911bdeb.el9.x 3.4 MB/s | 3.4 MB 00:01 2026-03-10T05:31:53.336 INFO:teuthology.orchestra.run.vm01.stdout:(45/143): mailcap-2.1.49-5.el9.noarch.rpm 1.0 MB/s | 33 kB 00:00 2026-03-10T05:31:53.339 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------------------------------------------------------- 2026-03-10T05:31:53.339 INFO:teuthology.orchestra.run.vm02.stdout:Total 18 MB/s | 212 MB 00:11 2026-03-10T05:31:53.394 INFO:teuthology.orchestra.run.vm01.stdout:(46/143): pciutils-3.7.0-7.el9.x86_64.rpm 1.6 MB/s | 93 kB 00:00 2026-03-10T05:31:53.500 INFO:teuthology.orchestra.run.vm01.stdout:(47/143): libquadmath-11.5.0-14.el9.x86_64.rpm 927 kB/s | 184 kB 00:00 2026-03-10T05:31:53.503 INFO:teuthology.orchestra.run.vm01.stdout:(48/143): python3-cffi-1.14.5-5.el9.x86_64.rpm 2.3 MB/s | 253 kB 00:00 2026-03-10T05:31:53.549 INFO:teuthology.orchestra.run.vm01.stdout:(49/143): python3-ply-3.11-14.el9.noarch.rpm 2.3 MB/s | 106 kB 00:00 2026-03-10T05:31:53.607 INFO:teuthology.orchestra.run.vm01.stdout:(50/143): python3-pycparser-2.20-6.el9.noarch.r 2.3 MB/s | 135 kB 00:00 2026-03-10T05:31:53.641 INFO:teuthology.orchestra.run.vm01.stdout:(51/143): python3-pyparsing-2.4.7-9.el9.noarch. 4.3 MB/s | 150 kB 00:00 2026-03-10T05:31:53.663 INFO:teuthology.orchestra.run.vm01.stdout:(52/143): python3-cryptography-36.0.1-5.el9.x86 7.6 MB/s | 1.2 MB 00:00 2026-03-10T05:31:53.691 INFO:teuthology.orchestra.run.vm01.stdout:(53/143): python3-requests-2.25.1-10.el9.noarch 2.5 MB/s | 126 kB 00:00 2026-03-10T05:31:53.730 INFO:teuthology.orchestra.run.vm05.stdout:(8/143): ceph-radosgw-19.2.3-678.ge911bdeb.el9. 13 MB/s | 11 MB 00:00 2026-03-10T05:31:53.730 INFO:teuthology.orchestra.run.vm01.stdout:(54/143): python3-urllib3-1.26.5-7.el9.noarch.r 3.2 MB/s | 218 kB 00:00 2026-03-10T05:31:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:53 vm05 ceph-mon[50927]: pgmap v129: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:53 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:53 vm01 ceph-mon[47941]: pgmap v129: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:53.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:53 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:53.850 INFO:teuthology.orchestra.run.vm01.stdout:(55/143): zip-3.0-35.el9.x86_64.rpm 2.2 MB/s | 266 kB 00:00 2026-03-10T05:31:53.851 INFO:teuthology.orchestra.run.vm05.stdout:(9/143): ceph-selinux-19.2.3-678.ge911bdeb.el9. 206 kB/s | 25 kB 00:00 2026-03-10T05:31:53.853 INFO:teuthology.orchestra.run.vm01.stdout:(56/143): unzip-6.0-59.el9.x86_64.rpm 1.1 MB/s | 182 kB 00:00 2026-03-10T05:31:53.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:53 vm02 ceph-mon[50473]: pgmap v129: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:53.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:53 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:31:53.915 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-10T05:31:53.965 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-10T05:31:53.965 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-10T05:31:54.249 INFO:teuthology.orchestra.run.vm01.stdout:(57/143): flexiblas-3.0.4-9.el9.x86_64.rpm 75 kB/s | 30 kB 00:00 2026-03-10T05:31:54.365 INFO:teuthology.orchestra.run.vm01.stdout:(58/143): boost-program-options-1.75.0-13.el9.x 202 kB/s | 104 kB 00:00 2026-03-10T05:31:54.437 INFO:teuthology.orchestra.run.vm01.stdout:(59/143): flexiblas-openblas-openmp-3.0.4-9.el9 207 kB/s | 15 kB 00:00 2026-03-10T05:31:54.569 INFO:teuthology.orchestra.run.vm01.stdout:(60/143): libnbd-1.20.3-4.el9.x86_64.rpm 1.2 MB/s | 164 kB 00:00 2026-03-10T05:31:54.657 INFO:teuthology.orchestra.run.vm01.stdout:(61/143): libpmemobj-1.12.1-1.el9.x86_64.rpm 1.8 MB/s | 160 kB 00:00 2026-03-10T05:31:54.681 INFO:teuthology.orchestra.run.vm01.stdout:(62/143): flexiblas-netlib-3.0.4-9.el9.x86_64.r 6.9 MB/s | 3.0 MB 00:00 2026-03-10T05:31:54.733 INFO:teuthology.orchestra.run.vm01.stdout:(63/143): librabbitmq-0.11.0-7.el9.x86_64.rpm 591 kB/s | 45 kB 00:00 2026-03-10T05:31:54.774 INFO:teuthology.orchestra.run.vm01.stdout:(64/143): librdkafka-1.6.1-102.el9.x86_64.rpm 7.0 MB/s | 662 kB 00:00 2026-03-10T05:31:54.782 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-10T05:31:54.782 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-10T05:31:54.833 INFO:teuthology.orchestra.run.vm01.stdout:(65/143): libstoragemgmt-1.10.1-1.el9.x86_64.rp 2.4 MB/s | 246 kB 00:00 2026-03-10T05:31:54.845 INFO:teuthology.orchestra.run.vm01.stdout:(66/143): libxslt-1.1.34-12.el9.x86_64.rpm 3.2 MB/s | 233 kB 00:00 2026-03-10T05:31:54.901 INFO:teuthology.orchestra.run.vm05.stdout:(10/143): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 7.6 MB/s | 17 MB 00:02 2026-03-10T05:31:54.941 INFO:teuthology.orchestra.run.vm01.stdout:(67/143): lttng-ust-2.12.0-6.el9.x86_64.rpm 2.7 MB/s | 292 kB 00:00 2026-03-10T05:31:54.952 INFO:teuthology.orchestra.run.vm01.stdout:(68/143): lua-5.4.4-4.el9.x86_64.rpm 1.7 MB/s | 188 kB 00:00 2026-03-10T05:31:54.975 INFO:teuthology.orchestra.run.vm01.stdout:(69/143): ceph-mgr-diskprediction-local-19.2.3- 2.2 MB/s | 7.4 MB 00:03 2026-03-10T05:31:55.025 INFO:teuthology.orchestra.run.vm01.stdout:(70/143): openblas-0.3.29-1.el9.x86_64.rpm 504 kB/s | 42 kB 00:00 2026-03-10T05:31:55.030 INFO:teuthology.orchestra.run.vm05.stdout:(11/143): libcephfs-devel-19.2.3-678.ge911bdeb. 260 kB/s | 34 kB 00:00 2026-03-10T05:31:55.259 INFO:teuthology.orchestra.run.vm05.stdout:(12/143): libcephfs2-19.2.3-678.ge911bdeb.el9.x 4.3 MB/s | 1.0 MB 00:00 2026-03-10T05:31:55.374 INFO:teuthology.orchestra.run.vm05.stdout:(13/143): libcephsqlite-19.2.3-678.ge911bdeb.el 1.4 MB/s | 163 kB 00:00 2026-03-10T05:31:55.420 INFO:teuthology.orchestra.run.vm01.stdout:(71/143): openblas-openmp-0.3.29-1.el9.x86_64.r 11 MB/s | 5.3 MB 00:00 2026-03-10T05:31:55.444 INFO:teuthology.orchestra.run.vm01.stdout:(72/143): python3-babel-2.9.1-2.el9.noarch.rpm 14 MB/s | 6.0 MB 00:00 2026-03-10T05:31:55.490 INFO:teuthology.orchestra.run.vm05.stdout:(14/143): librados-devel-19.2.3-678.ge911bdeb.e 1.1 MB/s | 127 kB 00:00 2026-03-10T05:31:55.493 INFO:teuthology.orchestra.run.vm01.stdout:(73/143): protobuf-3.14.0-17.el9.x86_64.rpm 1.9 MB/s | 1.0 MB 00:00 2026-03-10T05:31:55.546 INFO:teuthology.orchestra.run.vm01.stdout:(74/143): python3-devel-3.9.25-3.el9.x86_64.rpm 1.9 MB/s | 244 kB 00:00 2026-03-10T05:31:55.555 INFO:teuthology.orchestra.run.vm01.stdout:(75/143): python3-jinja2-2.11.3-8.el9.noarch.rp 2.2 MB/s | 249 kB 00:00 2026-03-10T05:31:55.579 INFO:teuthology.orchestra.run.vm01.stdout:(76/143): python3-jmespath-1.0.1-1.el9.noarch.r 555 kB/s | 48 kB 00:00 2026-03-10T05:31:55.610 INFO:teuthology.orchestra.run.vm05.stdout:(15/143): libradosstriper1-19.2.3-678.ge911bdeb 4.1 MB/s | 503 kB 00:00 2026-03-10T05:31:55.619 INFO:teuthology.orchestra.run.vm01.stdout:(77/143): python3-libstoragemgmt-1.10.1-1.el9.x 2.4 MB/s | 177 kB 00:00 2026-03-10T05:31:55.658 INFO:teuthology.orchestra.run.vm01.stdout:(78/143): python3-markupsafe-1.1.1-12.el9.x86_6 441 kB/s | 35 kB 00:00 2026-03-10T05:31:55.659 INFO:teuthology.orchestra.run.vm01.stdout:(79/143): python3-mako-1.1.4-6.el9.noarch.rpm 1.6 MB/s | 172 kB 00:00 2026-03-10T05:31:55.661 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-10T05:31:55.674 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/146 2026-03-10T05:31:55.687 INFO:teuthology.orchestra.run.vm02.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/146 2026-03-10T05:31:55.745 INFO:teuthology.orchestra.run.vm01.stdout:(80/143): python3-packaging-20.9-5.el9.noarch.r 901 kB/s | 77 kB 00:00 2026-03-10T05:31:55.760 INFO:teuthology.orchestra.run.vm01.stdout:(81/143): python3-numpy-f2py-1.23.5-2.el9.x86_6 4.2 MB/s | 442 kB 00:00 2026-03-10T05:31:55.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:55 vm05 ceph-mon[50927]: pgmap v130: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:55.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:55 vm01 ceph-mon[47941]: pgmap v130: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:55.851 INFO:teuthology.orchestra.run.vm01.stdout:(82/143): python3-pyasn1-0.4.8-7.el9.noarch.rpm 1.7 MB/s | 157 kB 00:00 2026-03-10T05:31:55.852 INFO:teuthology.orchestra.run.vm02.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/146 2026-03-10T05:31:55.853 INFO:teuthology.orchestra.run.vm01.stdout:(83/143): python3-protobuf-3.14.0-17.el9.noarch 2.4 MB/s | 267 kB 00:00 2026-03-10T05:31:55.854 INFO:teuthology.orchestra.run.vm02.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-10T05:31:55.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:55 vm02 ceph-mon[50473]: pgmap v130: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:55.914 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-10T05:31:55.915 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/146 2026-03-10T05:31:55.942 INFO:teuthology.orchestra.run.vm01.stdout:(84/143): python3-pyasn1-modules-0.4.8-7.el9.no 3.0 MB/s | 277 kB 00:00 2026-03-10T05:31:55.945 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/146 2026-03-10T05:31:55.954 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/146 2026-03-10T05:31:55.957 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/146 2026-03-10T05:31:55.960 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/146 2026-03-10T05:31:55.965 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/146 2026-03-10T05:31:55.975 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/146 2026-03-10T05:31:55.977 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-10T05:31:55.977 INFO:teuthology.orchestra.run.vm01.stdout:(85/143): python3-requests-oauthlib-1.3.0-12.el 433 kB/s | 54 kB 00:00 2026-03-10T05:31:56.014 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-10T05:31:56.015 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/146 2026-03-10T05:31:56.031 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/146 2026-03-10T05:31:56.067 INFO:teuthology.orchestra.run.vm02.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/146 2026-03-10T05:31:56.076 INFO:teuthology.orchestra.run.vm01.stdout:(86/143): python3-toml-0.10.2-6.el9.noarch.rpm 423 kB/s | 42 kB 00:00 2026-03-10T05:31:56.107 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/146 2026-03-10T05:31:56.113 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/146 2026-03-10T05:31:56.138 INFO:teuthology.orchestra.run.vm02.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/146 2026-03-10T05:31:56.146 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/146 2026-03-10T05:31:56.156 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 18/146 2026-03-10T05:31:56.163 INFO:teuthology.orchestra.run.vm02.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 19/146 2026-03-10T05:31:56.164 INFO:teuthology.orchestra.run.vm01.stdout:(87/143): qatlib-25.08.0-2.el9.x86_64.rpm 2.7 MB/s | 240 kB 00:00 2026-03-10T05:31:56.167 INFO:teuthology.orchestra.run.vm02.stdout: Installing : lua-5.4.4-4.el9.x86_64 20/146 2026-03-10T05:31:56.174 INFO:teuthology.orchestra.run.vm02.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 21/146 2026-03-10T05:31:56.203 INFO:teuthology.orchestra.run.vm02.stdout: Installing : unzip-6.0-59.el9.x86_64 22/146 2026-03-10T05:31:56.220 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 23/146 2026-03-10T05:31:56.224 INFO:teuthology.orchestra.run.vm01.stdout:(88/143): python3-numpy-1.23.5-2.el9.x86_64.rpm 10 MB/s | 6.1 MB 00:00 2026-03-10T05:31:56.226 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 24/146 2026-03-10T05:31:56.233 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 25/146 2026-03-10T05:31:56.236 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 26/146 2026-03-10T05:31:56.268 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 27/146 2026-03-10T05:31:56.275 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 28/146 2026-03-10T05:31:56.285 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 29/146 2026-03-10T05:31:56.292 INFO:teuthology.orchestra.run.vm01.stdout:(89/143): qatlib-service-25.08.0-2.el9.x86_64.r 291 kB/s | 37 kB 00:00 2026-03-10T05:31:56.300 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 30/146 2026-03-10T05:31:56.303 INFO:teuthology.orchestra.run.vm05.stdout:(16/143): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 7.8 MB/s | 5.4 MB 00:00 2026-03-10T05:31:56.309 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 31/146 2026-03-10T05:31:56.339 INFO:teuthology.orchestra.run.vm02.stdout: Installing : zip-3.0-35.el9.x86_64 32/146 2026-03-10T05:31:56.344 INFO:teuthology.orchestra.run.vm02.stdout: Installing : luarocks-3.9.2-5.el9.noarch 33/146 2026-03-10T05:31:56.353 INFO:teuthology.orchestra.run.vm02.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 34/146 2026-03-10T05:31:56.383 INFO:teuthology.orchestra.run.vm02.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 35/146 2026-03-10T05:31:56.414 INFO:teuthology.orchestra.run.vm01.stdout:(90/143): socat-1.7.4.1-8.el9.x86_64.rpm 2.4 MB/s | 303 kB 00:00 2026-03-10T05:31:56.417 INFO:teuthology.orchestra.run.vm05.stdout:(17/143): python3-ceph-argparse-19.2.3-678.ge91 397 kB/s | 45 kB 00:00 2026-03-10T05:31:56.418 INFO:teuthology.orchestra.run.vm01.stdout:(91/143): qatzip-libs-1.3.1-1.el9.x86_64.rpm 343 kB/s | 66 kB 00:00 2026-03-10T05:31:56.447 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 36/146 2026-03-10T05:31:56.466 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 37/146 2026-03-10T05:31:56.474 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rsa-4.9-2.el9.noarch 38/146 2026-03-10T05:31:56.483 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 39/146 2026-03-10T05:31:56.484 INFO:teuthology.orchestra.run.vm01.stdout:(92/143): xmlstarlet-1.6.1-20.el9.x86_64.rpm 920 kB/s | 64 kB 00:00 2026-03-10T05:31:56.487 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 40/146 2026-03-10T05:31:56.493 INFO:teuthology.orchestra.run.vm02.stdout: Upgrading : kpartx-0.8.7-45.el9.x86_64 41/146 2026-03-10T05:31:56.497 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/146 2026-03-10T05:31:56.513 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/146 2026-03-10T05:31:56.533 INFO:teuthology.orchestra.run.vm05.stdout:(18/143): python3-ceph-common-19.2.3-678.ge911b 1.2 MB/s | 142 kB 00:00 2026-03-10T05:31:56.539 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/146 2026-03-10T05:31:56.546 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/146 2026-03-10T05:31:56.553 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/146 2026-03-10T05:31:56.567 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/146 2026-03-10T05:31:56.579 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/146 2026-03-10T05:31:56.589 INFO:teuthology.orchestra.run.vm01.stdout:(93/143): lua-devel-5.4.4-4.el9.x86_64.rpm 130 kB/s | 22 kB 00:00 2026-03-10T05:31:56.591 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/146 2026-03-10T05:31:56.648 INFO:teuthology.orchestra.run.vm05.stdout:(19/143): python3-cephfs-19.2.3-678.ge911bdeb.e 1.4 MB/s | 165 kB 00:00 2026-03-10T05:31:56.654 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/146 2026-03-10T05:31:56.662 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/146 2026-03-10T05:31:56.671 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/146 2026-03-10T05:31:56.718 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/146 2026-03-10T05:31:56.767 INFO:teuthology.orchestra.run.vm05.stdout:(20/143): python3-rados-19.2.3-678.ge911bdeb.el 2.7 MB/s | 323 kB 00:00 2026-03-10T05:31:56.835 INFO:teuthology.orchestra.run.vm01.stdout:(94/143): protobuf-compiler-3.14.0-17.el9.x86_6 2.4 MB/s | 862 kB 00:00 2026-03-10T05:31:56.884 INFO:teuthology.orchestra.run.vm05.stdout:(21/143): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.5 MB/s | 303 kB 00:00 2026-03-10T05:31:56.978 INFO:teuthology.orchestra.run.vm01.stdout:(95/143): abseil-cpp-20211102.0-4.el9.x86_64.rp 1.4 MB/s | 551 kB 00:00 2026-03-10T05:31:56.998 INFO:teuthology.orchestra.run.vm05.stdout:(22/143): python3-rgw-19.2.3-678.ge911bdeb.el9. 872 kB/s | 100 kB 00:00 2026-03-10T05:31:57.052 INFO:teuthology.orchestra.run.vm01.stdout:(96/143): python3-scipy-1.9.3-2.el9.x86_64.rpm 17 MB/s | 19 MB 00:01 2026-03-10T05:31:57.053 INFO:teuthology.orchestra.run.vm01.stdout:(97/143): grpc-data-1.46.7-10.el9.noarch.rpm 256 kB/s | 19 kB 00:00 2026-03-10T05:31:57.100 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/146 2026-03-10T05:31:57.113 INFO:teuthology.orchestra.run.vm05.stdout:(23/143): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 743 kB/s | 85 kB 00:00 2026-03-10T05:31:57.115 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/146 2026-03-10T05:31:57.120 INFO:teuthology.orchestra.run.vm01.stdout:(98/143): libarrow-doc-9.0.0-15.el9.noarch.rpm 376 kB/s | 25 kB 00:00 2026-03-10T05:31:57.120 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/146 2026-03-10T05:31:57.128 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/146 2026-03-10T05:31:57.132 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/146 2026-03-10T05:31:57.140 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/146 2026-03-10T05:31:57.144 INFO:teuthology.orchestra.run.vm02.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/146 2026-03-10T05:31:57.146 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/146 2026-03-10T05:31:57.168 INFO:teuthology.orchestra.run.vm01.stdout:(99/143): liboath-2.6.12-1.el9.x86_64.rpm 1.0 MB/s | 49 kB 00:00 2026-03-10T05:31:57.177 INFO:teuthology.orchestra.run.vm02.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/146 2026-03-10T05:31:57.218 INFO:teuthology.orchestra.run.vm01.stdout:(100/143): libunwind-1.6.2-1.el9.x86_64.rpm 1.3 MB/s | 67 kB 00:00 2026-03-10T05:31:57.220 INFO:teuthology.orchestra.run.vm01.stdout:(101/143): gperftools-libs-2.9.1-3.el9.x86_64.r 800 kB/s | 308 kB 00:00 2026-03-10T05:31:57.231 INFO:teuthology.orchestra.run.vm02.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/146 2026-03-10T05:31:57.245 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/146 2026-03-10T05:31:57.253 INFO:teuthology.orchestra.run.vm02.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/146 2026-03-10T05:31:57.258 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/146 2026-03-10T05:31:57.265 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/146 2026-03-10T05:31:57.270 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/146 2026-03-10T05:31:57.272 INFO:teuthology.orchestra.run.vm01.stdout:(102/143): luarocks-3.9.2-5.el9.noarch.rpm 2.7 MB/s | 151 kB 00:00 2026-03-10T05:31:57.279 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/146 2026-03-10T05:31:57.284 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/146 2026-03-10T05:31:57.317 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/146 2026-03-10T05:31:57.330 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/146 2026-03-10T05:31:57.370 INFO:teuthology.orchestra.run.vm01.stdout:(103/143): python3-asyncssh-2.13.2-5.el9.noarch 5.5 MB/s | 548 kB 00:00 2026-03-10T05:31:57.373 INFO:teuthology.orchestra.run.vm01.stdout:(104/143): parquet-libs-9.0.0-15.el9.x86_64.rpm 5.4 MB/s | 838 kB 00:00 2026-03-10T05:31:57.410 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/146 2026-03-10T05:31:57.417 INFO:teuthology.orchestra.run.vm01.stdout:(105/143): python3-autocommand-2.2.2-8.el9.noar 630 kB/s | 29 kB 00:00 2026-03-10T05:31:57.421 INFO:teuthology.orchestra.run.vm01.stdout:(106/143): python3-backports-tarfile-1.2.0-1.el 1.2 MB/s | 60 kB 00:00 2026-03-10T05:31:57.444 INFO:teuthology.orchestra.run.vm05.stdout:(24/143): ceph-base-19.2.3-678.ge911bdeb.el9.x8 741 kB/s | 5.5 MB 00:07 2026-03-10T05:31:57.472 INFO:teuthology.orchestra.run.vm01.stdout:(107/143): python3-cachetools-4.2.4-1.el9.noarc 635 kB/s | 32 kB 00:00 2026-03-10T05:31:57.472 INFO:teuthology.orchestra.run.vm01.stdout:(108/143): python3-bcrypt-3.2.2-1.el9.x86_64.rp 785 kB/s | 43 kB 00:00 2026-03-10T05:31:57.472 INFO:teuthology.orchestra.run.vm05.stdout:(25/143): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 8.7 MB/s | 3.1 MB 00:00 2026-03-10T05:31:57.525 INFO:teuthology.orchestra.run.vm01.stdout:(109/143): python3-cheroot-10.0.1-4.el9.noarch. 3.2 MB/s | 173 kB 00:00 2026-03-10T05:31:57.526 INFO:teuthology.orchestra.run.vm01.stdout:(110/143): python3-certifi-2023.05.07-4.el9.noa 260 kB/s | 14 kB 00:00 2026-03-10T05:31:57.605 INFO:teuthology.orchestra.run.vm01.stdout:(111/143): python3-cherrypy-18.6.1-2.el9.noarch 4.4 MB/s | 358 kB 00:00 2026-03-10T05:31:57.606 INFO:teuthology.orchestra.run.vm05.stdout:(26/143): ceph-test-19.2.3-678.ge911bdeb.el9.x8 13 MB/s | 50 MB 00:03 2026-03-10T05:31:57.607 INFO:teuthology.orchestra.run.vm05.stdout:(27/143): ceph-grafana-dashboards-19.2.3-678.ge 231 kB/s | 31 kB 00:00 2026-03-10T05:31:57.608 INFO:teuthology.orchestra.run.vm01.stdout:(112/143): python3-google-auth-2.45.0-1.el9.noa 3.0 MB/s | 254 kB 00:00 2026-03-10T05:31:57.622 INFO:teuthology.orchestra.run.vm01.stdout:(113/143): libarrow-9.0.0-15.el9.x86_64.rpm 7.7 MB/s | 4.4 MB 00:00 2026-03-10T05:31:57.664 INFO:teuthology.orchestra.run.vm01.stdout:(114/143): python3-grpcio-tools-1.46.7-10.el9.x 2.6 MB/s | 144 kB 00:00 2026-03-10T05:31:57.669 INFO:teuthology.orchestra.run.vm01.stdout:(115/143): python3-jaraco-8.2.1-3.el9.noarch.rp 227 kB/s | 11 kB 00:00 2026-03-10T05:31:57.678 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/146 2026-03-10T05:31:57.708 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/146 2026-03-10T05:31:57.711 INFO:teuthology.orchestra.run.vm01.stdout:(116/143): python3-jaraco-classes-3.2.1-5.el9.n 377 kB/s | 18 kB 00:00 2026-03-10T05:31:57.715 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/146 2026-03-10T05:31:57.715 INFO:teuthology.orchestra.run.vm05.stdout:(28/143): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 633 kB/s | 171 kB 00:00 2026-03-10T05:31:57.716 INFO:teuthology.orchestra.run.vm01.stdout:(117/143): python3-jaraco-collections-3.0.0-8.e 497 kB/s | 23 kB 00:00 2026-03-10T05:31:57.721 INFO:teuthology.orchestra.run.vm05.stdout:(29/143): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.3 MB/s | 150 kB 00:00 2026-03-10T05:31:57.725 INFO:teuthology.orchestra.run.vm01.stdout:(118/143): python3-grpcio-1.46.7-10.el9.x86_64. 17 MB/s | 2.0 MB 00:00 2026-03-10T05:31:57.758 INFO:teuthology.orchestra.run.vm01.stdout:(119/143): python3-jaraco-context-6.0.1-3.el9.n 418 kB/s | 20 kB 00:00 2026-03-10T05:31:57.762 INFO:teuthology.orchestra.run.vm01.stdout:(120/143): python3-jaraco-functools-3.5.0-2.el9 419 kB/s | 19 kB 00:00 2026-03-10T05:31:57.773 INFO:teuthology.orchestra.run.vm01.stdout:(121/143): python3-jaraco-text-4.0.0-2.el9.noar 548 kB/s | 26 kB 00:00 2026-03-10T05:31:57.777 INFO:teuthology.orchestra.run.vm02.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/146 2026-03-10T05:31:57.780 INFO:teuthology.orchestra.run.vm02.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/146 2026-03-10T05:31:57.805 INFO:teuthology.orchestra.run.vm02.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/146 2026-03-10T05:31:57.809 INFO:teuthology.orchestra.run.vm01.stdout:(122/143): python3-logutils-0.3.5-21.el9.noarch 986 kB/s | 46 kB 00:00 2026-03-10T05:31:57.823 INFO:teuthology.orchestra.run.vm01.stdout:(123/143): python3-more-itertools-8.12.0-2.el9. 1.6 MB/s | 79 kB 00:00 2026-03-10T05:31:57.836 INFO:teuthology.orchestra.run.vm05.stdout:(30/143): ceph-mgr-modules-core-19.2.3-678.ge91 2.1 MB/s | 253 kB 00:00 2026-03-10T05:31:57.857 INFO:teuthology.orchestra.run.vm01.stdout:(124/143): python3-natsort-7.1.1-5.el9.noarch.r 1.2 MB/s | 58 kB 00:00 2026-03-10T05:31:57.863 INFO:teuthology.orchestra.run.vm01.stdout:(125/143): python3-kubernetes-26.1.0-3.el9.noar 9.7 MB/s | 1.0 MB 00:00 2026-03-10T05:31:57.879 INFO:teuthology.orchestra.run.vm01.stdout:(126/143): python3-pecan-1.4.2-3.el9.noarch.rpm 4.7 MB/s | 272 kB 00:00 2026-03-10T05:31:57.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:57 vm02 ceph-mon[50473]: pgmap v131: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:57.903 INFO:teuthology.orchestra.run.vm01.stdout:(127/143): python3-portend-3.1.0-2.el9.noarch.r 355 kB/s | 16 kB 00:00 2026-03-10T05:31:57.911 INFO:teuthology.orchestra.run.vm01.stdout:(128/143): python3-pyOpenSSL-21.0.0-1.el9.noarc 1.8 MB/s | 90 kB 00:00 2026-03-10T05:31:57.926 INFO:teuthology.orchestra.run.vm01.stdout:(129/143): python3-repoze-lru-0.7-16.el9.noarch 657 kB/s | 31 kB 00:00 2026-03-10T05:31:57.949 INFO:teuthology.orchestra.run.vm05.stdout:(31/143): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 436 kB/s | 49 kB 00:00 2026-03-10T05:31:57.955 INFO:teuthology.orchestra.run.vm01.stdout:(130/143): python3-routes-2.5.1-5.el9.noarch.rp 3.6 MB/s | 188 kB 00:00 2026-03-10T05:31:57.960 INFO:teuthology.orchestra.run.vm01.stdout:(131/143): python3-rsa-4.9-2.el9.noarch.rpm 1.2 MB/s | 59 kB 00:00 2026-03-10T05:31:57.973 INFO:teuthology.orchestra.run.vm01.stdout:(132/143): python3-tempora-5.0.0-2.el9.noarch.r 762 kB/s | 36 kB 00:00 2026-03-10T05:31:58.002 INFO:teuthology.orchestra.run.vm01.stdout:(133/143): python3-typing-extensions-4.15.0-1.e 1.8 MB/s | 86 kB 00:00 2026-03-10T05:31:58.013 INFO:teuthology.orchestra.run.vm01.stdout:(134/143): python3-webob-1.8.8-2.el9.noarch.rpm 4.2 MB/s | 230 kB 00:00 2026-03-10T05:31:58.021 INFO:teuthology.orchestra.run.vm01.stdout:(135/143): python3-websocket-client-1.2.3-2.el9 1.8 MB/s | 90 kB 00:00 2026-03-10T05:31:58.053 INFO:teuthology.orchestra.run.vm01.stdout:(136/143): python3-werkzeug-2.0.3-3.el9.1.noarc 8.1 MB/s | 427 kB 00:00 2026-03-10T05:31:58.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:57 vm05 ceph-mon[50927]: pgmap v131: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:58.060 INFO:teuthology.orchestra.run.vm01.stdout:(137/143): python3-xmltodict-0.12.0-15.el9.noar 478 kB/s | 22 kB 00:00 2026-03-10T05:31:58.062 INFO:teuthology.orchestra.run.vm05.stdout:(32/143): ceph-prometheus-alerts-19.2.3-678.ge9 149 kB/s | 17 kB 00:00 2026-03-10T05:31:58.068 INFO:teuthology.orchestra.run.vm01.stdout:(138/143): python3-zc-lockfile-2.0-10.el9.noarc 429 kB/s | 20 kB 00:00 2026-03-10T05:31:58.079 INFO:teuthology.orchestra.run.vm05.stdout:(33/143): ceph-mgr-dashboard-19.2.3-678.ge911bd 8.1 MB/s | 3.8 MB 00:00 2026-03-10T05:31:58.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:57 vm01 ceph-mon[47941]: pgmap v131: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:31:58.101 INFO:teuthology.orchestra.run.vm01.stdout:(139/143): re2-20211101-20.el9.x86_64.rpm 3.9 MB/s | 191 kB 00:00 2026-03-10T05:31:58.138 INFO:teuthology.orchestra.run.vm01.stdout:(140/143): thrift-0.15.0-4.el9.x86_64.rpm 20 MB/s | 1.6 MB 00:00 2026-03-10T05:31:58.178 INFO:teuthology.orchestra.run.vm05.stdout:(34/143): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.5 MB/s | 299 kB 00:00 2026-03-10T05:31:58.178 INFO:teuthology.orchestra.run.vm01.stdout:(141/143): kpartx-0.8.7-45.el9.x86_64.rpm 1.2 MB/s | 49 kB 00:00 2026-03-10T05:31:58.187 INFO:teuthology.orchestra.run.vm02.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/146 2026-03-10T05:31:58.201 INFO:teuthology.orchestra.run.vm05.stdout:(35/143): cephadm-19.2.3-678.ge911bdeb.el9.noar 6.2 MB/s | 769 kB 00:00 2026-03-10T05:31:58.276 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/146 2026-03-10T05:31:58.324 INFO:teuthology.orchestra.run.vm05.stdout:(36/143): cryptsetup-2.8.1-3.el9.x86_64.rpm 2.4 MB/s | 351 kB 00:00 2026-03-10T05:31:58.325 INFO:teuthology.orchestra.run.vm05.stdout:(37/143): device-mapper-multipath-0.8.7-45.el9. 1.2 MB/s | 156 kB 00:00 2026-03-10T05:31:58.395 INFO:teuthology.orchestra.run.vm05.stdout:(38/143): device-mapper-multipath-libs-0.8.7-45 4.0 MB/s | 289 kB 00:00 2026-03-10T05:31:58.413 INFO:teuthology.orchestra.run.vm05.stdout:(39/143): iscsi-initiator-utils-6.2.1.11-0.git4 4.4 MB/s | 392 kB 00:00 2026-03-10T05:31:58.445 INFO:teuthology.orchestra.run.vm05.stdout:(40/143): isns-utils-libs-0.101-4.el9.x86_64.rp 3.1 MB/s | 101 kB 00:00 2026-03-10T05:31:58.447 INFO:teuthology.orchestra.run.vm05.stdout:(41/143): iscsi-initiator-utils-iscsiuio-6.2.1. 1.5 MB/s | 81 kB 00:00 2026-03-10T05:31:58.481 INFO:teuthology.orchestra.run.vm05.stdout:(42/143): libconfig-1.7.2-9.el9.x86_64.rpm 2.1 MB/s | 72 kB 00:00 2026-03-10T05:31:58.484 INFO:teuthology.orchestra.run.vm05.stdout:(43/143): ledmon-libs-1.1.0-3.el9.x86_64.rpm 1.0 MB/s | 40 kB 00:00 2026-03-10T05:31:58.548 INFO:teuthology.orchestra.run.vm05.stdout:(44/143): libquadmath-11.5.0-14.el9.x86_64.rpm 2.9 MB/s | 184 kB 00:00 2026-03-10T05:31:58.565 INFO:teuthology.orchestra.run.vm05.stdout:(45/143): libgfortran-11.5.0-14.el9.x86_64.rpm 9.3 MB/s | 794 kB 00:00 2026-03-10T05:31:58.566 INFO:teuthology.orchestra.run.vm05.stdout:(46/143): mailcap-2.1.49-5.el9.noarch.rpm 1.8 MB/s | 33 kB 00:00 2026-03-10T05:31:58.604 INFO:teuthology.orchestra.run.vm05.stdout:(47/143): pciutils-3.7.0-7.el9.x86_64.rpm 2.4 MB/s | 93 kB 00:00 2026-03-10T05:31:58.622 INFO:teuthology.orchestra.run.vm05.stdout:(48/143): python3-cffi-1.14.5-5.el9.x86_64.rpm 4.4 MB/s | 253 kB 00:00 2026-03-10T05:31:58.648 INFO:teuthology.orchestra.run.vm05.stdout:(49/143): python3-ply-3.11-14.el9.noarch.rpm 4.0 MB/s | 106 kB 00:00 2026-03-10T05:31:58.697 INFO:teuthology.orchestra.run.vm05.stdout:(50/143): python3-cryptography-36.0.1-5.el9.x86 13 MB/s | 1.2 MB 00:00 2026-03-10T05:31:58.699 INFO:teuthology.orchestra.run.vm05.stdout:(51/143): python3-pycparser-2.20-6.el9.noarch.r 2.6 MB/s | 135 kB 00:00 2026-03-10T05:31:58.732 INFO:teuthology.orchestra.run.vm05.stdout:(52/143): python3-requests-2.25.1-10.el9.noarch 3.8 MB/s | 126 kB 00:00 2026-03-10T05:31:58.740 INFO:teuthology.orchestra.run.vm05.stdout:(53/143): python3-pyparsing-2.4.7-9.el9.noarch. 3.4 MB/s | 150 kB 00:00 2026-03-10T05:31:58.775 INFO:teuthology.orchestra.run.vm05.stdout:(54/143): python3-urllib3-1.26.5-7.el9.noarch.r 4.9 MB/s | 218 kB 00:00 2026-03-10T05:31:58.802 INFO:teuthology.orchestra.run.vm05.stdout:(55/143): unzip-6.0-59.el9.x86_64.rpm 2.9 MB/s | 182 kB 00:00 2026-03-10T05:31:58.842 INFO:teuthology.orchestra.run.vm05.stdout:(56/143): zip-3.0-35.el9.x86_64.rpm 3.9 MB/s | 266 kB 00:00 2026-03-10T05:31:58.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:31:58 vm02 ceph-mon[50473]: pgmap v132: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:59.034 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/146 2026-03-10T05:31:59.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:31:58 vm05 ceph-mon[50927]: pgmap v132: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:59.061 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/146 2026-03-10T05:31:59.068 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/146 2026-03-10T05:31:59.072 INFO:teuthology.orchestra.run.vm02.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/146 2026-03-10T05:31:59.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:31:58 vm01 ceph-mon[47941]: pgmap v132: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:31:59.103 INFO:teuthology.orchestra.run.vm05.stdout:(57/143): flexiblas-3.0.4-9.el9.x86_64.rpm 114 kB/s | 30 kB 00:00 2026-03-10T05:31:59.221 INFO:teuthology.orchestra.run.vm05.stdout:(58/143): boost-program-options-1.75.0-13.el9.x 248 kB/s | 104 kB 00:00 2026-03-10T05:31:59.223 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/146 2026-03-10T05:31:59.226 INFO:teuthology.orchestra.run.vm02.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/146 2026-03-10T05:31:59.256 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/146 2026-03-10T05:31:59.259 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/146 2026-03-10T05:31:59.267 INFO:teuthology.orchestra.run.vm02.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/146 2026-03-10T05:31:59.324 INFO:teuthology.orchestra.run.vm05.stdout:(59/143): flexiblas-openblas-openmp-3.0.4-9.el9 144 kB/s | 15 kB 00:00 2026-03-10T05:31:59.494 INFO:teuthology.orchestra.run.vm05.stdout:(60/143): libnbd-1.20.3-4.el9.x86_64.rpm 964 kB/s | 164 kB 00:00 2026-03-10T05:31:59.510 INFO:teuthology.orchestra.run.vm02.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/146 2026-03-10T05:31:59.513 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/146 2026-03-10T05:31:59.531 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/146 2026-03-10T05:31:59.532 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/146 2026-03-10T05:31:59.602 INFO:teuthology.orchestra.run.vm05.stdout:(61/143): libpmemobj-1.12.1-1.el9.x86_64.rpm 1.5 MB/s | 160 kB 00:00 2026-03-10T05:31:59.668 INFO:teuthology.orchestra.run.vm05.stdout:(62/143): flexiblas-netlib-3.0.4-9.el9.x86_64.r 5.3 MB/s | 3.0 MB 00:00 2026-03-10T05:31:59.700 INFO:teuthology.orchestra.run.vm05.stdout:(63/143): librabbitmq-0.11.0-7.el9.x86_64.rpm 465 kB/s | 45 kB 00:00 2026-03-10T05:31:59.780 INFO:teuthology.orchestra.run.vm05.stdout:(64/143): librdkafka-1.6.1-102.el9.x86_64.rpm 5.8 MB/s | 662 kB 00:00 2026-03-10T05:31:59.827 INFO:teuthology.orchestra.run.vm05.stdout:(65/143): libstoragemgmt-1.10.1-1.el9.x86_64.rp 1.9 MB/s | 246 kB 00:00 2026-03-10T05:31:59.880 INFO:teuthology.orchestra.run.vm05.stdout:(66/143): libxslt-1.1.34-12.el9.x86_64.rpm 2.3 MB/s | 233 kB 00:00 2026-03-10T05:31:59.958 INFO:teuthology.orchestra.run.vm05.stdout:(67/143): lttng-ust-2.12.0-6.el9.x86_64.rpm 2.2 MB/s | 292 kB 00:00 2026-03-10T05:31:59.997 INFO:teuthology.orchestra.run.vm05.stdout:(68/143): lua-5.4.4-4.el9.x86_64.rpm 1.6 MB/s | 188 kB 00:00 2026-03-10T05:31:59.997 INFO:teuthology.orchestra.run.vm01.stdout:(142/143): librbd1-19.2.3-678.ge911bdeb.el9.x86 1.7 MB/s | 3.2 MB 00:01 2026-03-10T05:32:00.051 INFO:teuthology.orchestra.run.vm05.stdout:(69/143): openblas-0.3.29-1.el9.x86_64.rpm 452 kB/s | 42 kB 00:00 2026-03-10T05:32:00.273 INFO:teuthology.orchestra.run.vm05.stdout:(70/143): protobuf-3.14.0-17.el9.x86_64.rpm 4.5 MB/s | 1.0 MB 00:00 2026-03-10T05:32:00.302 INFO:teuthology.orchestra.run.vm05.stdout:(71/143): openblas-openmp-0.3.29-1.el9.x86_64.r 17 MB/s | 5.3 MB 00:00 2026-03-10T05:32:00.402 INFO:teuthology.orchestra.run.vm05.stdout:(72/143): python3-devel-3.9.25-3.el9.x86_64.rpm 2.4 MB/s | 244 kB 00:00 2026-03-10T05:32:00.515 INFO:teuthology.orchestra.run.vm05.stdout:(73/143): python3-jinja2-2.11.3-8.el9.noarch.rp 2.2 MB/s | 249 kB 00:00 2026-03-10T05:32:00.598 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-10T05:32:00.603 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-10T05:32:00.611 INFO:teuthology.orchestra.run.vm05.stdout:(74/143): python3-jmespath-1.0.1-1.el9.noarch.r 495 kB/s | 48 kB 00:00 2026-03-10T05:32:00.622 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-10T05:32:00.634 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 94/146 2026-03-10T05:32:00.642 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-packaging-20.9-5.el9.noarch 95/146 2026-03-10T05:32:00.647 INFO:teuthology.orchestra.run.vm01.stdout:(143/143): librados2-19.2.3-678.ge911bdeb.el9.x 1.3 MB/s | 3.4 MB 00:02 2026-03-10T05:32:00.651 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-10T05:32:00.651 INFO:teuthology.orchestra.run.vm01.stdout:Total 12 MB/s | 212 MB 00:17 2026-03-10T05:32:00.661 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-ply-3.11-14.el9.noarch 96/146 2026-03-10T05:32:00.662 INFO:teuthology.orchestra.run.vm05.stdout:(75/143): python3-babel-2.9.1-2.el9.noarch.rpm 15 MB/s | 6.0 MB 00:00 2026-03-10T05:32:00.682 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 97/146 2026-03-10T05:32:00.729 INFO:teuthology.orchestra.run.vm05.stdout:(76/143): python3-libstoragemgmt-1.10.1-1.el9.x 1.5 MB/s | 177 kB 00:00 2026-03-10T05:32:00.770 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 98/146 2026-03-10T05:32:00.779 INFO:teuthology.orchestra.run.vm05.stdout:(77/143): python3-mako-1.1.4-6.el9.noarch.rpm 1.4 MB/s | 172 kB 00:00 2026-03-10T05:32:00.783 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 99/146 2026-03-10T05:32:00.812 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 100/146 2026-03-10T05:32:00.830 INFO:teuthology.orchestra.run.vm05.stdout:(78/143): python3-markupsafe-1.1.1-12.el9.x86_6 345 kB/s | 35 kB 00:00 2026-03-10T05:32:00.848 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 101/146 2026-03-10T05:32:00.908 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 102/146 2026-03-10T05:32:00.920 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 103/146 2026-03-10T05:32:00.927 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 104/146 2026-03-10T05:32:00.932 INFO:teuthology.orchestra.run.vm02.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 105/146 2026-03-10T05:32:00.936 INFO:teuthology.orchestra.run.vm02.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 106/146 2026-03-10T05:32:00.938 INFO:teuthology.orchestra.run.vm02.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 107/146 2026-03-10T05:32:00.955 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 107/146 2026-03-10T05:32:00.958 INFO:teuthology.orchestra.run.vm05.stdout:(79/143): python3-numpy-f2py-1.23.5-2.el9.x86_6 3.4 MB/s | 442 kB 00:00 2026-03-10T05:32:01.055 INFO:teuthology.orchestra.run.vm05.stdout:(80/143): python3-packaging-20.9-5.el9.noarch.r 806 kB/s | 77 kB 00:00 2026-03-10T05:32:01.159 INFO:teuthology.orchestra.run.vm05.stdout:(81/143): python3-protobuf-3.14.0-17.el9.noarch 2.5 MB/s | 267 kB 00:00 2026-03-10T05:32:01.175 INFO:teuthology.orchestra.run.vm05.stdout:(82/143): python3-numpy-1.23.5-2.el9.x86_64.rpm 16 MB/s | 6.1 MB 00:00 2026-03-10T05:32:01.247 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-10T05:32:01.250 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:00 vm01 ceph-mon[47941]: pgmap v133: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:01.252 INFO:teuthology.orchestra.run.vm05.stdout:(83/143): python3-pyasn1-0.4.8-7.el9.noarch.rpm 1.6 MB/s | 157 kB 00:00 2026-03-10T05:32:01.265 INFO:teuthology.orchestra.run.vm02.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 108/146 2026-03-10T05:32:01.286 INFO:teuthology.orchestra.run.vm05.stdout:(84/143): python3-pyasn1-modules-0.4.8-7.el9.no 2.4 MB/s | 277 kB 00:00 2026-03-10T05:32:01.288 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 109/146 2026-03-10T05:32:01.297 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-10T05:32:01.297 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-10T05:32:01.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:00 vm05 ceph-mon[50927]: pgmap v133: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:01.328 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 109/146 2026-03-10T05:32:01.328 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-10T05:32:01.328 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:01.331 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:00 vm02 ceph-mon[50473]: pgmap v133: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:01.332 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 110/146 2026-03-10T05:32:01.343 INFO:teuthology.orchestra.run.vm05.stdout:(85/143): python3-requests-oauthlib-1.3.0-12.el 596 kB/s | 54 kB 00:00 2026-03-10T05:32:01.429 INFO:teuthology.orchestra.run.vm05.stdout:(86/143): python3-toml-0.10.2-6.el9.noarch.rpm 487 kB/s | 42 kB 00:00 2026-03-10T05:32:01.536 INFO:teuthology.orchestra.run.vm05.stdout:(87/143): qatlib-25.08.0-2.el9.x86_64.rpm 2.2 MB/s | 240 kB 00:00 2026-03-10T05:32:01.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:32:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:32:01.623 INFO:teuthology.orchestra.run.vm05.stdout:(88/143): qatlib-service-25.08.0-2.el9.x86_64.r 429 kB/s | 37 kB 00:00 2026-03-10T05:32:01.735 INFO:teuthology.orchestra.run.vm05.stdout:(89/143): qatzip-libs-1.3.1-1.el9.x86_64.rpm 595 kB/s | 66 kB 00:00 2026-03-10T05:32:01.857 INFO:teuthology.orchestra.run.vm05.stdout:(90/143): socat-1.7.4.1-8.el9.x86_64.rpm 2.4 MB/s | 303 kB 00:00 2026-03-10T05:32:01.966 INFO:teuthology.orchestra.run.vm05.stdout:(91/143): xmlstarlet-1.6.1-20.el9.x86_64.rpm 585 kB/s | 64 kB 00:00 2026-03-10T05:32:02.064 INFO:teuthology.orchestra.run.vm05.stdout:(92/143): lua-devel-5.4.4-4.el9.x86_64.rpm 227 kB/s | 22 kB 00:00 2026-03-10T05:32:02.125 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-10T05:32:02.126 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-10T05:32:02.248 INFO:teuthology.orchestra.run.vm05.stdout:(93/143): python3-scipy-1.9.3-2.el9.x86_64.rpm 20 MB/s | 19 MB 00:00 2026-03-10T05:32:02.285 INFO:teuthology.orchestra.run.vm05.stdout:(94/143): protobuf-compiler-3.14.0-17.el9.x86_6 3.8 MB/s | 862 kB 00:00 2026-03-10T05:32:02.411 INFO:teuthology.orchestra.run.vm05.stdout:(95/143): abseil-cpp-20211102.0-4.el9.x86_64.rp 3.3 MB/s | 551 kB 00:00 2026-03-10T05:32:02.428 INFO:teuthology.orchestra.run.vm05.stdout:(96/143): gperftools-libs-2.9.1-3.el9.x86_64.rp 2.1 MB/s | 308 kB 00:00 2026-03-10T05:32:02.433 INFO:teuthology.orchestra.run.vm05.stdout:(97/143): grpc-data-1.46.7-10.el9.noarch.rpm 865 kB/s | 19 kB 00:00 2026-03-10T05:32:02.466 INFO:teuthology.orchestra.run.vm05.stdout:(98/143): ceph-mgr-diskprediction-local-19.2.3- 1.6 MB/s | 7.4 MB 00:04 2026-03-10T05:32:02.467 INFO:teuthology.orchestra.run.vm05.stdout:(99/143): libarrow-doc-9.0.0-15.el9.noarch.rpm 739 kB/s | 25 kB 00:00 2026-03-10T05:32:02.491 INFO:teuthology.orchestra.run.vm05.stdout:(100/143): libunwind-1.6.2-1.el9.x86_64.rpm 2.8 MB/s | 67 kB 00:00 2026-03-10T05:32:02.521 INFO:teuthology.orchestra.run.vm05.stdout:(101/143): luarocks-3.9.2-5.el9.noarch.rpm 4.9 MB/s | 151 kB 00:00 2026-03-10T05:32:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:02 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:02 vm05 ceph-mon[50927]: from='client.? 192.168.123.102:0/36773786' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-10T05:32:02.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:32:02 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:32:02.562 INFO:teuthology.orchestra.run.vm05.stdout:(102/143): libarrow-9.0.0-15.el9.x86_64.rpm 33 MB/s | 4.4 MB 00:00 2026-03-10T05:32:02.562 INFO:teuthology.orchestra.run.vm05.stdout:(103/143): liboath-2.6.12-1.el9.x86_64.rpm 509 kB/s | 49 kB 00:00 2026-03-10T05:32:02.577 INFO:teuthology.orchestra.run.vm05.stdout:(104/143): parquet-libs-9.0.0-15.el9.x86_64.rpm 15 MB/s | 838 kB 00:00 2026-03-10T05:32:02.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:02 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:02.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:02 vm01 ceph-mon[47941]: from='client.? 192.168.123.102:0/36773786' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-10T05:32:02.591 INFO:teuthology.orchestra.run.vm05.stdout:(105/143): python3-asyncssh-2.13.2-5.el9.noarch 19 MB/s | 548 kB 00:00 2026-03-10T05:32:02.592 INFO:teuthology.orchestra.run.vm05.stdout:(106/143): python3-autocommand-2.2.2-8.el9.noar 1.0 MB/s | 29 kB 00:00 2026-03-10T05:32:02.601 INFO:teuthology.orchestra.run.vm05.stdout:(107/143): python3-backports-tarfile-1.2.0-1.el 2.5 MB/s | 60 kB 00:00 2026-03-10T05:32:02.613 INFO:teuthology.orchestra.run.vm05.stdout:(108/143): python3-bcrypt-3.2.2-1.el9.x86_64.rp 1.9 MB/s | 43 kB 00:00 2026-03-10T05:32:02.623 INFO:teuthology.orchestra.run.vm05.stdout:(109/143): python3-certifi-2023.05.07-4.el9.noa 635 kB/s | 14 kB 00:00 2026-03-10T05:32:02.624 INFO:teuthology.orchestra.run.vm05.stdout:(110/143): python3-cachetools-4.2.4-1.el9.noarc 994 kB/s | 32 kB 00:00 2026-03-10T05:32:02.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:02 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:02.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:02 vm02 ceph-mon[50473]: from='client.? 192.168.123.102:0/36773786' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-10T05:32:02.697 INFO:teuthology.orchestra.run.vm05.stdout:(111/143): python3-cheroot-10.0.1-4.el9.noarch. 2.0 MB/s | 173 kB 00:00 2026-03-10T05:32:02.701 INFO:teuthology.orchestra.run.vm05.stdout:(112/143): python3-cherrypy-18.6.1-2.el9.noarch 4.5 MB/s | 358 kB 00:00 2026-03-10T05:32:02.759 INFO:teuthology.orchestra.run.vm05.stdout:(113/143): python3-google-auth-2.45.0-1.el9.noa 1.8 MB/s | 254 kB 00:00 2026-03-10T05:32:02.782 INFO:teuthology.orchestra.run.vm05.stdout:(114/143): python3-jaraco-8.2.1-3.el9.noarch.rp 462 kB/s | 11 kB 00:00 2026-03-10T05:32:02.784 INFO:teuthology.orchestra.run.vm05.stdout:(115/143): python3-grpcio-tools-1.46.7-10.el9.x 1.7 MB/s | 144 kB 00:00 2026-03-10T05:32:02.798 INFO:teuthology.orchestra.run.vm05.stdout:(116/143): python3-grpcio-1.46.7-10.el9.x86_64. 20 MB/s | 2.0 MB 00:00 2026-03-10T05:32:02.807 INFO:teuthology.orchestra.run.vm05.stdout:(117/143): python3-jaraco-classes-3.2.1-5.el9.n 718 kB/s | 18 kB 00:00 2026-03-10T05:32:02.808 INFO:teuthology.orchestra.run.vm05.stdout:(118/143): python3-jaraco-collections-3.0.0-8.e 986 kB/s | 23 kB 00:00 2026-03-10T05:32:02.821 INFO:teuthology.orchestra.run.vm05.stdout:(119/143): python3-jaraco-context-6.0.1-3.el9.n 863 kB/s | 20 kB 00:00 2026-03-10T05:32:02.831 INFO:teuthology.orchestra.run.vm05.stdout:(120/143): python3-jaraco-functools-3.5.0-2.el9 836 kB/s | 19 kB 00:00 2026-03-10T05:32:02.833 INFO:teuthology.orchestra.run.vm05.stdout:(121/143): python3-jaraco-text-4.0.0-2.el9.noar 1.0 MB/s | 26 kB 00:00 2026-03-10T05:32:02.858 INFO:teuthology.orchestra.run.vm05.stdout:(122/143): python3-kubernetes-26.1.0-3.el9.noar 28 MB/s | 1.0 MB 00:00 2026-03-10T05:32:02.859 INFO:teuthology.orchestra.run.vm05.stdout:(123/143): python3-logutils-0.3.5-21.el9.noarch 1.6 MB/s | 46 kB 00:00 2026-03-10T05:32:02.866 INFO:teuthology.orchestra.run.vm05.stdout:(124/143): python3-more-itertools-8.12.0-2.el9. 2.3 MB/s | 79 kB 00:00 2026-03-10T05:32:02.881 INFO:teuthology.orchestra.run.vm05.stdout:(125/143): python3-natsort-7.1.1-5.el9.noarch.r 2.5 MB/s | 58 kB 00:00 2026-03-10T05:32:02.887 INFO:teuthology.orchestra.run.vm05.stdout:(126/143): python3-pecan-1.4.2-3.el9.noarch.rpm 9.5 MB/s | 272 kB 00:00 2026-03-10T05:32:02.891 INFO:teuthology.orchestra.run.vm05.stdout:(127/143): python3-portend-3.1.0-2.el9.noarch.r 676 kB/s | 16 kB 00:00 2026-03-10T05:32:02.908 INFO:teuthology.orchestra.run.vm05.stdout:(128/143): python3-pyOpenSSL-21.0.0-1.el9.noarc 3.3 MB/s | 90 kB 00:00 2026-03-10T05:32:02.912 INFO:teuthology.orchestra.run.vm05.stdout:(129/143): python3-repoze-lru-0.7-16.el9.noarch 1.2 MB/s | 31 kB 00:00 2026-03-10T05:32:02.935 INFO:teuthology.orchestra.run.vm05.stdout:(130/143): python3-rsa-4.9-2.el9.noarch.rpm 2.1 MB/s | 59 kB 00:00 2026-03-10T05:32:02.936 INFO:teuthology.orchestra.run.vm05.stdout:(131/143): python3-tempora-5.0.0-2.el9.noarch.r 1.4 MB/s | 36 kB 00:00 2026-03-10T05:32:02.940 INFO:teuthology.orchestra.run.vm05.stdout:(132/143): python3-routes-2.5.1-5.el9.noarch.rp 3.7 MB/s | 188 kB 00:00 2026-03-10T05:32:02.959 INFO:teuthology.orchestra.run.vm05.stdout:(133/143): python3-typing-extensions-4.15.0-1.e 3.5 MB/s | 86 kB 00:00 2026-03-10T05:32:02.962 INFO:teuthology.orchestra.run.vm05.stdout:(134/143): python3-webob-1.8.8-2.el9.noarch.rpm 8.6 MB/s | 230 kB 00:00 2026-03-10T05:32:02.966 INFO:teuthology.orchestra.run.vm05.stdout:(135/143): python3-websocket-client-1.2.3-2.el9 3.3 MB/s | 90 kB 00:00 2026-03-10T05:32:02.988 INFO:teuthology.orchestra.run.vm05.stdout:(136/143): python3-werkzeug-2.0.3-3.el9.1.noarc 15 MB/s | 427 kB 00:00 2026-03-10T05:32:02.988 INFO:teuthology.orchestra.run.vm05.stdout:(137/143): python3-xmltodict-0.12.0-15.el9.noar 848 kB/s | 22 kB 00:00 2026-03-10T05:32:02.993 INFO:teuthology.orchestra.run.vm05.stdout:(138/143): python3-zc-lockfile-2.0-10.el9.noarc 766 kB/s | 20 kB 00:00 2026-03-10T05:32:03.013 INFO:teuthology.orchestra.run.vm05.stdout:(139/143): re2-20211101-20.el9.x86_64.rpm 7.5 MB/s | 191 kB 00:00 2026-03-10T05:32:03.036 INFO:teuthology.orchestra.run.vm05.stdout:(140/143): thrift-0.15.0-4.el9.x86_64.rpm 34 MB/s | 1.6 MB 00:00 2026-03-10T05:32:03.099 INFO:teuthology.orchestra.run.vm05.stdout:(141/143): kpartx-0.8.7-45.el9.x86_64.rpm 779 kB/s | 49 kB 00:00 2026-03-10T05:32:03.180 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-10T05:32:03.193 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/146 2026-03-10T05:32:03.205 INFO:teuthology.orchestra.run.vm01.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/146 2026-03-10T05:32:03.371 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/146 2026-03-10T05:32:03.373 INFO:teuthology.orchestra.run.vm01.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-10T05:32:03.431 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-10T05:32:03.433 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/146 2026-03-10T05:32:03.462 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/146 2026-03-10T05:32:03.471 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/146 2026-03-10T05:32:03.475 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/146 2026-03-10T05:32:03.477 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/146 2026-03-10T05:32:03.482 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/146 2026-03-10T05:32:03.494 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/146 2026-03-10T05:32:03.547 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-10T05:32:03.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:03 vm05 ceph-mon[50927]: pgmap v134: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:03.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:03 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:03 vm01 ceph-mon[47941]: pgmap v134: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:03 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:03.590 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-10T05:32:03.592 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/146 2026-03-10T05:32:03.608 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/146 2026-03-10T05:32:03.640 INFO:teuthology.orchestra.run.vm01.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/146 2026-03-10T05:32:03.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:03 vm02 ceph-mon[50473]: pgmap v134: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:03.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:03 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:03.679 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/146 2026-03-10T05:32:03.685 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/146 2026-03-10T05:32:03.711 INFO:teuthology.orchestra.run.vm01.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/146 2026-03-10T05:32:03.721 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/146 2026-03-10T05:32:03.732 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 18/146 2026-03-10T05:32:03.739 INFO:teuthology.orchestra.run.vm01.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 19/146 2026-03-10T05:32:03.742 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lua-5.4.4-4.el9.x86_64 20/146 2026-03-10T05:32:03.747 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 21/146 2026-03-10T05:32:03.774 INFO:teuthology.orchestra.run.vm01.stdout: Installing : unzip-6.0-59.el9.x86_64 22/146 2026-03-10T05:32:03.790 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 23/146 2026-03-10T05:32:03.794 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 24/146 2026-03-10T05:32:03.801 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 25/146 2026-03-10T05:32:03.803 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 26/146 2026-03-10T05:32:03.835 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 27/146 2026-03-10T05:32:03.840 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 28/146 2026-03-10T05:32:03.850 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 29/146 2026-03-10T05:32:03.863 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 30/146 2026-03-10T05:32:03.871 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 31/146 2026-03-10T05:32:03.902 INFO:teuthology.orchestra.run.vm01.stdout: Installing : zip-3.0-35.el9.x86_64 32/146 2026-03-10T05:32:03.906 INFO:teuthology.orchestra.run.vm01.stdout: Installing : luarocks-3.9.2-5.el9.noarch 33/146 2026-03-10T05:32:03.914 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 34/146 2026-03-10T05:32:03.945 INFO:teuthology.orchestra.run.vm01.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 35/146 2026-03-10T05:32:04.011 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 36/146 2026-03-10T05:32:04.018 INFO:teuthology.orchestra.run.vm05.stdout:(142/143): librados2-19.2.3-678.ge911bdeb.el9.x 3.3 MB/s | 3.4 MB 00:01 2026-03-10T05:32:04.027 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 37/146 2026-03-10T05:32:04.034 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rsa-4.9-2.el9.noarch 38/146 2026-03-10T05:32:04.044 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 39/146 2026-03-10T05:32:04.048 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 40/146 2026-03-10T05:32:04.053 INFO:teuthology.orchestra.run.vm01.stdout: Upgrading : kpartx-0.8.7-45.el9.x86_64 41/146 2026-03-10T05:32:04.058 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/146 2026-03-10T05:32:04.076 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/146 2026-03-10T05:32:04.103 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/146 2026-03-10T05:32:04.109 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/146 2026-03-10T05:32:04.116 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/146 2026-03-10T05:32:04.130 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/146 2026-03-10T05:32:04.141 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/146 2026-03-10T05:32:04.152 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/146 2026-03-10T05:32:04.166 INFO:teuthology.orchestra.run.vm05.stdout:(143/143): librbd1-19.2.3-678.ge911bdeb.el9.x86 2.7 MB/s | 3.2 MB 00:01 2026-03-10T05:32:04.168 INFO:teuthology.orchestra.run.vm05.stdout:-------------------------------------------------------------------------------- 2026-03-10T05:32:04.169 INFO:teuthology.orchestra.run.vm05.stdout:Total 14 MB/s | 212 MB 00:15 2026-03-10T05:32:04.220 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/146 2026-03-10T05:32:04.227 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/146 2026-03-10T05:32:04.237 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/146 2026-03-10T05:32:04.286 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/146 2026-03-10T05:32:04.664 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/146 2026-03-10T05:32:04.678 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/146 2026-03-10T05:32:04.684 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/146 2026-03-10T05:32:04.692 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/146 2026-03-10T05:32:04.696 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/146 2026-03-10T05:32:04.704 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/146 2026-03-10T05:32:04.708 INFO:teuthology.orchestra.run.vm01.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/146 2026-03-10T05:32:04.710 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/146 2026-03-10T05:32:04.742 INFO:teuthology.orchestra.run.vm01.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/146 2026-03-10T05:32:04.762 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-10T05:32:04.795 INFO:teuthology.orchestra.run.vm01.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/146 2026-03-10T05:32:04.808 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/146 2026-03-10T05:32:04.811 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-10T05:32:04.811 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-10T05:32:04.816 INFO:teuthology.orchestra.run.vm01.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/146 2026-03-10T05:32:04.822 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/146 2026-03-10T05:32:04.829 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/146 2026-03-10T05:32:04.835 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/146 2026-03-10T05:32:04.844 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/146 2026-03-10T05:32:04.849 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/146 2026-03-10T05:32:04.882 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/146 2026-03-10T05:32:04.896 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/146 2026-03-10T05:32:04.938 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/146 2026-03-10T05:32:05.221 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/146 2026-03-10T05:32:05.253 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/146 2026-03-10T05:32:05.260 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/146 2026-03-10T05:32:05.325 INFO:teuthology.orchestra.run.vm01.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/146 2026-03-10T05:32:05.328 INFO:teuthology.orchestra.run.vm01.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/146 2026-03-10T05:32:05.354 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/146 2026-03-10T05:32:05.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:05 vm05 ceph-mon[50927]: pgmap v135: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:05.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:05 vm01 ceph-mon[47941]: pgmap v135: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:05.634 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-10T05:32:05.634 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-10T05:32:05.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:05 vm02 ceph-mon[50473]: pgmap v135: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:05.746 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/146 2026-03-10T05:32:05.842 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/146 2026-03-10T05:32:06.528 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-10T05:32:06.580 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/146 2026-03-10T05:32:06.607 INFO:teuthology.orchestra.run.vm05.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/146 2026-03-10T05:32:06.664 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/146 2026-03-10T05:32:06.792 INFO:teuthology.orchestra.run.vm05.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/146 2026-03-10T05:32:06.794 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/146 2026-03-10T05:32:06.815 INFO:teuthology.orchestra.run.vm05.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-10T05:32:06.875 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/146 2026-03-10T05:32:06.959 INFO:teuthology.orchestra.run.vm01.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/146 2026-03-10T05:32:07.011 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-10T05:32:07.065 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/146 2026-03-10T05:32:07.146 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/146 2026-03-10T05:32:07.155 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/146 2026-03-10T05:32:07.212 INFO:teuthology.orchestra.run.vm05.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/146 2026-03-10T05:32:07.214 INFO:teuthology.orchestra.run.vm05.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/146 2026-03-10T05:32:07.219 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/146 2026-03-10T05:32:07.225 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/146 2026-03-10T05:32:07.227 INFO:teuthology.orchestra.run.vm01.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/146 2026-03-10T05:32:07.231 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/146 2026-03-10T05:32:07.234 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-10T05:32:07.260 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/146 2026-03-10T05:32:07.265 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/146 2026-03-10T05:32:07.270 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-10T05:32:07.271 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/146 2026-03-10T05:32:07.272 INFO:teuthology.orchestra.run.vm01.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/146 2026-03-10T05:32:07.283 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/146 2026-03-10T05:32:07.318 INFO:teuthology.orchestra.run.vm05.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/146 2026-03-10T05:32:07.354 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/146 2026-03-10T05:32:07.358 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/146 2026-03-10T05:32:07.382 INFO:teuthology.orchestra.run.vm05.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/146 2026-03-10T05:32:07.390 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/146 2026-03-10T05:32:07.400 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 18/146 2026-03-10T05:32:07.406 INFO:teuthology.orchestra.run.vm05.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 19/146 2026-03-10T05:32:07.410 INFO:teuthology.orchestra.run.vm05.stdout: Installing : lua-5.4.4-4.el9.x86_64 20/146 2026-03-10T05:32:07.415 INFO:teuthology.orchestra.run.vm05.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 21/146 2026-03-10T05:32:07.441 INFO:teuthology.orchestra.run.vm05.stdout: Installing : unzip-6.0-59.el9.x86_64 22/146 2026-03-10T05:32:07.457 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 23/146 2026-03-10T05:32:07.462 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 24/146 2026-03-10T05:32:07.469 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 25/146 2026-03-10T05:32:07.471 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 26/146 2026-03-10T05:32:07.500 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 27/146 2026-03-10T05:32:07.506 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 28/146 2026-03-10T05:32:07.516 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 29/146 2026-03-10T05:32:07.527 INFO:teuthology.orchestra.run.vm01.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/146 2026-03-10T05:32:07.529 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/146 2026-03-10T05:32:07.530 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 30/146 2026-03-10T05:32:07.538 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 31/146 2026-03-10T05:32:07.551 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/146 2026-03-10T05:32:07.553 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/146 2026-03-10T05:32:07.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:07 vm05 ceph-mon[50927]: pgmap v136: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:07.566 INFO:teuthology.orchestra.run.vm05.stdout: Installing : zip-3.0-35.el9.x86_64 32/146 2026-03-10T05:32:07.569 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:07 vm01 ceph-mon[47941]: pgmap v136: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:07.571 INFO:teuthology.orchestra.run.vm05.stdout: Installing : luarocks-3.9.2-5.el9.noarch 33/146 2026-03-10T05:32:07.583 INFO:teuthology.orchestra.run.vm05.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 34/146 2026-03-10T05:32:07.612 INFO:teuthology.orchestra.run.vm05.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 35/146 2026-03-10T05:32:07.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:07 vm02 ceph-mon[50473]: pgmap v136: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:07.673 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 36/146 2026-03-10T05:32:07.689 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 37/146 2026-03-10T05:32:07.696 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-rsa-4.9-2.el9.noarch 38/146 2026-03-10T05:32:07.705 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 39/146 2026-03-10T05:32:07.709 INFO:teuthology.orchestra.run.vm05.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 40/146 2026-03-10T05:32:07.716 INFO:teuthology.orchestra.run.vm05.stdout: Upgrading : kpartx-0.8.7-45.el9.x86_64 41/146 2026-03-10T05:32:07.721 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/146 2026-03-10T05:32:07.740 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/146 2026-03-10T05:32:07.764 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/146 2026-03-10T05:32:07.771 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/146 2026-03-10T05:32:07.774 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 110/146 2026-03-10T05:32:07.774 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /sys 2026-03-10T05:32:07.774 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /proc 2026-03-10T05:32:07.774 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /mnt 2026-03-10T05:32:07.774 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /var/tmp 2026-03-10T05:32:07.774 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /home 2026-03-10T05:32:07.774 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /root 2026-03-10T05:32:07.774 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /tmp 2026-03-10T05:32:07.774 INFO:teuthology.orchestra.run.vm02.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.b.service-pid 2026-03-10T05:32:07.774 INFO:teuthology.orchestra.run.vm02.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.2.service-pid 2026-03-10T05:32:07.774 INFO:teuthology.orchestra.run.vm02.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.3.service-pid 2026-03-10T05:32:07.774 INFO:teuthology.orchestra.run.vm02.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.4.service-pid 2026-03-10T05:32:07.774 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:07.777 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/146 2026-03-10T05:32:07.791 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/146 2026-03-10T05:32:07.802 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/146 2026-03-10T05:32:07.813 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/146 2026-03-10T05:32:07.877 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/146 2026-03-10T05:32:07.884 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/146 2026-03-10T05:32:07.894 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/146 2026-03-10T05:32:07.901 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 111/146 2026-03-10T05:32:07.926 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 111/146 2026-03-10T05:32:07.926 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:07.926 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T05:32:07.926 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T05:32:07.926 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T05:32:07.926 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:07.942 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/146 2026-03-10T05:32:08.159 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 112/146 2026-03-10T05:32:08.182 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 112/146 2026-03-10T05:32:08.182 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:08.182 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T05:32:08.182 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T05:32:08.182 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T05:32:08.182 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:08.189 INFO:teuthology.orchestra.run.vm02.stdout: Installing : mailcap-2.1.49-5.el9.noarch 113/146 2026-03-10T05:32:08.191 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 114/146 2026-03-10T05:32:08.209 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-10T05:32:08.209 INFO:teuthology.orchestra.run.vm02.stdout:Creating group 'qat' with GID 994. 2026-03-10T05:32:08.209 INFO:teuthology.orchestra.run.vm02.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-10T05:32:08.209 INFO:teuthology.orchestra.run.vm02.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-10T05:32:08.209 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:08.218 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-10T05:32:08.249 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-10T05:32:08.249 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-10T05:32:08.249 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:08.265 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 116/146 2026-03-10T05:32:08.269 INFO:teuthology.orchestra.run.vm02.stdout: Installing : isns-utils-libs-0.101-4.el9.x86_64 117/146 2026-03-10T05:32:08.270 INFO:teuthology.orchestra.run.vm02.stdout: Installing : iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 118/146 2026-03-10T05:32:08.286 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 118/146 2026-03-10T05:32:08.286 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/sockets.target.wants/iscsiuio.socket → /usr/lib/systemd/system/iscsiuio.socket. 2026-03-10T05:32:08.286 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:08.299 INFO:teuthology.orchestra.run.vm02.stdout: Installing : iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 119/146 2026-03-10T05:32:08.316 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 119/146 2026-03-10T05:32:08.316 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/iscsi-starter.service → /usr/lib/systemd/system/iscsi-starter.service. 2026-03-10T05:32:08.316 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/sockets.target.wants/iscsid.socket → /usr/lib/systemd/system/iscsid.socket. 2026-03-10T05:32:08.316 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/iscsi-onboot.service → /usr/lib/systemd/system/iscsi-onboot.service. 2026-03-10T05:32:08.316 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:08.316 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/146 2026-03-10T05:32:08.332 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/146 2026-03-10T05:32:08.337 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/146 2026-03-10T05:32:08.345 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/146 2026-03-10T05:32:08.349 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/146 2026-03-10T05:32:08.356 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/146 2026-03-10T05:32:08.357 INFO:teuthology.orchestra.run.vm02.stdout: Installing : device-mapper-multipath-libs-0.8.7-45.el9.x86_64 120/146 2026-03-10T05:32:08.360 INFO:teuthology.orchestra.run.vm05.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/146 2026-03-10T05:32:08.362 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/146 2026-03-10T05:32:08.391 INFO:teuthology.orchestra.run.vm05.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/146 2026-03-10T05:32:08.435 INFO:teuthology.orchestra.run.vm02.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 121/146 2026-03-10T05:32:08.439 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 122/146 2026-03-10T05:32:08.441 INFO:teuthology.orchestra.run.vm05.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/146 2026-03-10T05:32:08.454 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/146 2026-03-10T05:32:08.454 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 122/146 2026-03-10T05:32:08.454 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:08.454 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T05:32:08.454 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:08.461 INFO:teuthology.orchestra.run.vm05.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/146 2026-03-10T05:32:08.466 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/146 2026-03-10T05:32:08.473 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/146 2026-03-10T05:32:08.478 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/146 2026-03-10T05:32:08.487 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/146 2026-03-10T05:32:08.492 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/146 2026-03-10T05:32:08.524 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/146 2026-03-10T05:32:08.538 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/146 2026-03-10T05:32:08.584 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/146 2026-03-10T05:32:08.689 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-10T05:32:08.721 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-10T05:32:08.743 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-10T05:32:08.756 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 94/146 2026-03-10T05:32:08.764 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-packaging-20.9-5.el9.noarch 95/146 2026-03-10T05:32:08.780 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ply-3.11-14.el9.noarch 96/146 2026-03-10T05:32:08.800 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 97/146 2026-03-10T05:32:08.848 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/146 2026-03-10T05:32:08.879 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/146 2026-03-10T05:32:08.886 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/146 2026-03-10T05:32:08.888 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 98/146 2026-03-10T05:32:08.901 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 99/146 2026-03-10T05:32:08.929 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 100/146 2026-03-10T05:32:08.948 INFO:teuthology.orchestra.run.vm05.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/146 2026-03-10T05:32:08.952 INFO:teuthology.orchestra.run.vm05.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/146 2026-03-10T05:32:08.965 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 101/146 2026-03-10T05:32:08.976 INFO:teuthology.orchestra.run.vm05.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/146 2026-03-10T05:32:09.025 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 102/146 2026-03-10T05:32:09.033 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 103/146 2026-03-10T05:32:09.039 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 104/146 2026-03-10T05:32:09.044 INFO:teuthology.orchestra.run.vm01.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 105/146 2026-03-10T05:32:09.048 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 106/146 2026-03-10T05:32:09.050 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 107/146 2026-03-10T05:32:09.069 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 107/146 2026-03-10T05:32:09.237 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 123/146 2026-03-10T05:32:09.264 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 123/146 2026-03-10T05:32:09.264 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:09.264 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T05:32:09.264 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T05:32:09.264 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T05:32:09.264 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:09.319 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 124/146 2026-03-10T05:32:09.322 INFO:teuthology.orchestra.run.vm02.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 124/146 2026-03-10T05:32:09.328 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 125/146 2026-03-10T05:32:09.350 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 126/146 2026-03-10T05:32:09.352 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 127/146 2026-03-10T05:32:09.371 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 108/146 2026-03-10T05:32:09.376 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 109/146 2026-03-10T05:32:09.381 INFO:teuthology.orchestra.run.vm05.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/146 2026-03-10T05:32:09.438 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 109/146 2026-03-10T05:32:09.438 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-10T05:32:09.438 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:09.439 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 110/146 2026-03-10T05:32:09.475 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/146 2026-03-10T05:32:09.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:09 vm05 ceph-mon[50927]: pgmap v137: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:09.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:09 vm01 ceph-mon[47941]: pgmap v137: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:09.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:09 vm02 ceph-mon[50473]: pgmap v137: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:09.892 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 127/146 2026-03-10T05:32:09.898 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 128/146 2026-03-10T05:32:10.261 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/146 2026-03-10T05:32:10.285 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/146 2026-03-10T05:32:10.292 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/146 2026-03-10T05:32:10.297 INFO:teuthology.orchestra.run.vm05.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/146 2026-03-10T05:32:10.420 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 128/146 2026-03-10T05:32:10.455 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/146 2026-03-10T05:32:10.486 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 129/146 2026-03-10T05:32:10.486 INFO:teuthology.orchestra.run.vm05.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/146 2026-03-10T05:32:10.520 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/146 2026-03-10T05:32:10.524 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/146 2026-03-10T05:32:10.531 INFO:teuthology.orchestra.run.vm05.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/146 2026-03-10T05:32:10.551 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 129/146 2026-03-10T05:32:10.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:10 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3669031388' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-10T05:32:10.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:10 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3669031388' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-10T05:32:10.608 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 130/146 2026-03-10T05:32:10.610 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 131/146 2026-03-10T05:32:10.633 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:10 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3669031388' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-10T05:32:10.635 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 131/146 2026-03-10T05:32:10.635 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:10.635 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T05:32:10.635 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T05:32:10.635 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T05:32:10.635 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:10.648 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 132/146 2026-03-10T05:32:10.661 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 132/146 2026-03-10T05:32:10.669 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 133/146 2026-03-10T05:32:10.672 INFO:teuthology.orchestra.run.vm02.stdout: Installing : device-mapper-multipath-0.8.7-45.el9.x86_64 134/146 2026-03-10T05:32:10.687 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: device-mapper-multipath-0.8.7-45.el9.x86_64 134/146 2026-03-10T05:32:10.687 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/multipathd.service → /usr/lib/systemd/system/multipathd.service. 2026-03-10T05:32:10.687 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/sockets.target.wants/multipathd.socket → /usr/lib/systemd/system/multipathd.socket. 2026-03-10T05:32:10.687 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:10.791 INFO:teuthology.orchestra.run.vm05.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/146 2026-03-10T05:32:10.794 INFO:teuthology.orchestra.run.vm05.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/146 2026-03-10T05:32:10.814 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/146 2026-03-10T05:32:10.816 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/146 2026-03-10T05:32:11.196 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 135/146 2026-03-10T05:32:11.219 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 135/146 2026-03-10T05:32:11.219 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:11.219 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T05:32:11.219 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T05:32:11.219 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T05:32:11.219 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:11.229 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 136/146 2026-03-10T05:32:11.252 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 136/146 2026-03-10T05:32:11.252 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:11.252 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T05:32:11.252 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:11.405 INFO:teuthology.orchestra.run.vm02.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 137/146 2026-03-10T05:32:11.428 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 137/146 2026-03-10T05:32:11.428 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:11.428 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T05:32:11.428 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T05:32:11.428 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T05:32:11.428 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:11.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:11 vm05 ceph-mon[50927]: pgmap v138: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:11.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:32:11 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:32:11.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:11 vm01 ceph-mon[47941]: pgmap v138: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:11.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:11 vm02 ceph-mon[50473]: pgmap v138: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:11.957 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-10T05:32:11.961 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-10T05:32:11.984 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/146 2026-03-10T05:32:11.996 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 94/146 2026-03-10T05:32:12.006 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-packaging-20.9-5.el9.noarch 95/146 2026-03-10T05:32:12.025 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-ply-3.11-14.el9.noarch 96/146 2026-03-10T05:32:12.046 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 97/146 2026-03-10T05:32:12.139 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 98/146 2026-03-10T05:32:12.155 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 99/146 2026-03-10T05:32:12.186 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 100/146 2026-03-10T05:32:12.226 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 101/146 2026-03-10T05:32:12.291 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 102/146 2026-03-10T05:32:12.302 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 103/146 2026-03-10T05:32:12.308 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 104/146 2026-03-10T05:32:12.315 INFO:teuthology.orchestra.run.vm05.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 105/146 2026-03-10T05:32:12.320 INFO:teuthology.orchestra.run.vm05.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 106/146 2026-03-10T05:32:12.322 INFO:teuthology.orchestra.run.vm05.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 107/146 2026-03-10T05:32:12.342 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 107/146 2026-03-10T05:32:12.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:12 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:12.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:32:12 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:32:12.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:12 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:12.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:12 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:12.666 INFO:teuthology.orchestra.run.vm05.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 108/146 2026-03-10T05:32:12.687 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 109/146 2026-03-10T05:32:12.729 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 109/146 2026-03-10T05:32:12.729 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-10T05:32:12.729 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:12.733 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 110/146 2026-03-10T05:32:13.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:13 vm05 ceph-mon[50927]: pgmap v139: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:13.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:13 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:13.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:13 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3596891858' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-10T05:32:13.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:13 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:32:13.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:13 vm01 ceph-mon[47941]: pgmap v139: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:13.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:13 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:13.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:13 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3596891858' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-10T05:32:13.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:13 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:32:13.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:13 vm02 ceph-mon[50473]: pgmap v139: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:13.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:13 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:13.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:13 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3596891858' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-10T05:32:13.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:13 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:32:14.178 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 138/146 2026-03-10T05:32:14.193 INFO:teuthology.orchestra.run.vm02.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 139/146 2026-03-10T05:32:14.198 INFO:teuthology.orchestra.run.vm02.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 140/146 2026-03-10T05:32:14.258 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 141/146 2026-03-10T05:32:14.322 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 142/146 2026-03-10T05:32:14.342 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 143/146 2026-03-10T05:32:14.342 INFO:teuthology.orchestra.run.vm02.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-10T05:32:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-10T05:32:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 145/146 2026-03-10T05:32:14.383 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 145/146 2026-03-10T05:32:14.403 INFO:teuthology.orchestra.run.vm02.stdout: Cleanup : kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-10T05:32:14.491 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:14 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:32:14.491 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:14 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:32:14.491 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:14 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:32:14.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:14 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:32:14.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:14 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:32:14.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:14 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:32:14.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:32:14.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:32:14.854 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:32:16.787 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-10T05:32:16.787 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/146 2026-03-10T05:32:16.787 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/146 2026-03-10T05:32:16.787 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/146 2026-03-10T05:32:16.787 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-10T05:32:16.787 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/146 2026-03-10T05:32:16.787 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/146 2026-03-10T05:32:16.787 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/146 2026-03-10T05:32:16.787 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/146 2026-03-10T05:32:16.787 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/146 2026-03-10T05:32:16.787 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/146 2026-03-10T05:32:16.787 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/146 2026-03-10T05:32:16.788 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : device-mapper-multipath-0.8.7-45.el9.x86_64 38/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : device-mapper-multipath-libs-0.8.7-45.el9.x86_64 39/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 40/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 41/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : isns-utils-libs-0.101-4.el9.x86_64 42/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 43/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 44/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 45/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 46/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 47/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 48/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 49/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 50/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ply-3.11-14.el9.noarch 51/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 52/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 53/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 54/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 55/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : unzip-6.0-59.el9.x86_64 56/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : zip-3.0-35.el9.x86_64 57/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 58/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 59/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 60/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 61/146 2026-03-10T05:32:16.789 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 62/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 63/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 64/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 65/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 66/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 67/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 68/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-5.4.4-4.el9.x86_64 69/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 70/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 71/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 72/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 73/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 74/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 75/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 76/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 77/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 78/146 2026-03-10T05:32:16.790 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 79/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 80/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 81/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 82/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 83/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 84/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 85/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 87/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 88/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 89/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 90/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 91/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 92/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 93/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 94/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 95/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 96/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 97/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 98/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 99/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 100/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 101/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 102/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 103/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 104/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 105/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 106/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 107/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 108/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 109/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 110/146 2026-03-10T05:32:16.791 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 111/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 112/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 113/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 114/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 115/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 116/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 117/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 118/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 119/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 120/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 121/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 122/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 123/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 124/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 125/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 126/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 127/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 128/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 129/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 130/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 131/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 132/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 133/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 134/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 135/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 136/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 137/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 138/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : re2-1:20211101-20.el9.x86_64 139/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 140/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 141/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 142/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 143/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-10T05:32:16.792 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : kpartx-0.8.7-45.el9.x86_64 145/146 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout:Upgraded: 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: kpartx-0.8.7-45.el9.x86_64 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout:Installed: 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:16.969 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: device-mapper-multipath-0.8.7-45.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: device-mapper-multipath-libs-0.8.7-45.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3e853.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: isns-utils-libs-0.101-4.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.970 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyparsing-2.4.7-9.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T05:32:16.971 INFO:teuthology.orchestra.run.vm02.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T05:32:16.972 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T05:32:16.972 INFO:teuthology.orchestra.run.vm02.stdout: zip-3.0-35.el9.x86_64 2026-03-10T05:32:16.972 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:16.972 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:32:17.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:16 vm05 ceph-mon[50927]: pgmap v140: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:17.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:16 vm01 ceph-mon[47941]: pgmap v140: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:17.133 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:16 vm02 ceph-mon[50473]: pgmap v140: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:17.156 DEBUG:teuthology.parallel:result is None 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 110/146 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /sys 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /proc 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /mnt 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /var/tmp 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /home 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /root 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /tmp 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-sidecar@iscsi.iscsi.a:tcmu.service-pid 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@iscsi.iscsi.a.service-pid 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mgr.x.service-pid 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.a.service-pid 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.0.service-pid 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.1.service-pid 2026-03-10T05:32:17.261 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:17.387 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 111/146 2026-03-10T05:32:17.409 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 111/146 2026-03-10T05:32:17.409 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:17.409 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T05:32:17.409 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T05:32:17.409 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T05:32:17.409 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:17.639 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 112/146 2026-03-10T05:32:17.660 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 112/146 2026-03-10T05:32:17.660 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:17.660 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T05:32:17.660 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T05:32:17.660 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T05:32:17.660 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:17.668 INFO:teuthology.orchestra.run.vm01.stdout: Installing : mailcap-2.1.49-5.el9.noarch 113/146 2026-03-10T05:32:17.670 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 114/146 2026-03-10T05:32:17.688 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-10T05:32:17.688 INFO:teuthology.orchestra.run.vm01.stdout:Creating group 'qat' with GID 994. 2026-03-10T05:32:17.688 INFO:teuthology.orchestra.run.vm01.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-10T05:32:17.688 INFO:teuthology.orchestra.run.vm01.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-10T05:32:17.688 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:17.714 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-10T05:32:17.823 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-10T05:32:17.823 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-10T05:32:17.823 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:17.862 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 116/146 2026-03-10T05:32:17.931 INFO:teuthology.orchestra.run.vm01.stdout: Installing : isns-utils-libs-0.101-4.el9.x86_64 117/146 2026-03-10T05:32:17.982 INFO:teuthology.orchestra.run.vm01.stdout: Installing : iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 118/146 2026-03-10T05:32:17.999 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 118/146 2026-03-10T05:32:17.999 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/sockets.target.wants/iscsiuio.socket → /usr/lib/systemd/system/iscsiuio.socket. 2026-03-10T05:32:17.999 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:18.013 INFO:teuthology.orchestra.run.vm01.stdout: Installing : iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 119/146 2026-03-10T05:32:18.032 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 119/146 2026-03-10T05:32:18.032 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/iscsi-starter.service → /usr/lib/systemd/system/iscsi-starter.service. 2026-03-10T05:32:18.032 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/sockets.target.wants/iscsid.socket → /usr/lib/systemd/system/iscsid.socket. 2026-03-10T05:32:18.033 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/iscsi-onboot.service → /usr/lib/systemd/system/iscsi-onboot.service. 2026-03-10T05:32:18.033 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:18.171 INFO:teuthology.orchestra.run.vm01.stdout: Installing : device-mapper-multipath-libs-0.8.7-45.el9.x86_64 120/146 2026-03-10T05:32:18.306 INFO:teuthology.orchestra.run.vm01.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 121/146 2026-03-10T05:32:18.330 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 122/146 2026-03-10T05:32:18.345 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 122/146 2026-03-10T05:32:18.345 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:18.345 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T05:32:18.345 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:19.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:18 vm05 ceph-mon[50927]: pgmap v141: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:19.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:18 vm01 ceph-mon[47941]: pgmap v141: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:19.144 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 123/146 2026-03-10T05:32:19.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:18 vm02 ceph-mon[50473]: pgmap v141: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:19.169 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 123/146 2026-03-10T05:32:19.169 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:19.169 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T05:32:19.169 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T05:32:19.169 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T05:32:19.169 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 110/146 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /sys 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /proc 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /mnt 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /var/tmp 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /home 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /root 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /tmp 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-sidecar@iscsi.iscsi.b:tcmu.service-pid 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@iscsi.iscsi.b.service-pid 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.c.service-pid 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.5.service-pid 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.6.service-pid 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout:Warning no default label for /run/ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.7.service-pid 2026-03-10T05:32:19.225 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:19.226 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 124/146 2026-03-10T05:32:19.228 INFO:teuthology.orchestra.run.vm01.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 124/146 2026-03-10T05:32:19.234 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 125/146 2026-03-10T05:32:19.257 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 126/146 2026-03-10T05:32:19.259 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 127/146 2026-03-10T05:32:19.345 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 111/146 2026-03-10T05:32:19.368 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 111/146 2026-03-10T05:32:19.368 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:19.368 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T05:32:19.368 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T05:32:19.368 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T05:32:19.368 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:19.596 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 112/146 2026-03-10T05:32:19.618 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 112/146 2026-03-10T05:32:19.618 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:19.618 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T05:32:19.618 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T05:32:19.618 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T05:32:19.618 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:19.635 INFO:teuthology.orchestra.run.vm05.stdout: Installing : mailcap-2.1.49-5.el9.noarch 113/146 2026-03-10T05:32:19.640 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 114/146 2026-03-10T05:32:19.652 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:19 vm05 ceph-mon[50927]: pgmap v142: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:19.659 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-10T05:32:19.659 INFO:teuthology.orchestra.run.vm05.stdout:Creating group 'qat' with GID 994. 2026-03-10T05:32:19.659 INFO:teuthology.orchestra.run.vm05.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-10T05:32:19.659 INFO:teuthology.orchestra.run.vm05.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-10T05:32:19.659 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:19.668 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-10T05:32:19.695 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 115/146 2026-03-10T05:32:19.695 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-10T05:32:19.695 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:19.712 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 116/146 2026-03-10T05:32:19.715 INFO:teuthology.orchestra.run.vm05.stdout: Installing : isns-utils-libs-0.101-4.el9.x86_64 117/146 2026-03-10T05:32:19.717 INFO:teuthology.orchestra.run.vm05.stdout: Installing : iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 118/146 2026-03-10T05:32:19.733 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 118/146 2026-03-10T05:32:19.733 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/sockets.target.wants/iscsiuio.socket → /usr/lib/systemd/system/iscsiuio.socket. 2026-03-10T05:32:19.733 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:19.744 INFO:teuthology.orchestra.run.vm05.stdout: Installing : iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 119/146 2026-03-10T05:32:19.761 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 119/146 2026-03-10T05:32:19.761 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/iscsi-starter.service → /usr/lib/systemd/system/iscsi-starter.service. 2026-03-10T05:32:19.761 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/sockets.target.wants/iscsid.socket → /usr/lib/systemd/system/iscsid.socket. 2026-03-10T05:32:19.761 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/iscsi-onboot.service → /usr/lib/systemd/system/iscsi-onboot.service. 2026-03-10T05:32:19.761 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:19.801 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 127/146 2026-03-10T05:32:19.802 INFO:teuthology.orchestra.run.vm05.stdout: Installing : device-mapper-multipath-libs-0.8.7-45.el9.x86_64 120/146 2026-03-10T05:32:19.807 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 128/146 2026-03-10T05:32:19.876 INFO:teuthology.orchestra.run.vm05.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 121/146 2026-03-10T05:32:19.880 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 122/146 2026-03-10T05:32:19.893 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 122/146 2026-03-10T05:32:19.893 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:19.893 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T05:32:19.893 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:20.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:19 vm01 ceph-mon[47941]: pgmap v142: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:20.147 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:19 vm02 ceph-mon[50473]: pgmap v142: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:20.300 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 128/146 2026-03-10T05:32:20.303 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 129/146 2026-03-10T05:32:20.361 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 129/146 2026-03-10T05:32:20.416 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 130/146 2026-03-10T05:32:20.418 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 131/146 2026-03-10T05:32:20.439 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 131/146 2026-03-10T05:32:20.440 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:20.440 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T05:32:20.440 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T05:32:20.440 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T05:32:20.440 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:20.452 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 132/146 2026-03-10T05:32:20.465 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 132/146 2026-03-10T05:32:20.472 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 133/146 2026-03-10T05:32:20.475 INFO:teuthology.orchestra.run.vm01.stdout: Installing : device-mapper-multipath-0.8.7-45.el9.x86_64 134/146 2026-03-10T05:32:20.490 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: device-mapper-multipath-0.8.7-45.el9.x86_64 134/146 2026-03-10T05:32:20.491 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/multipathd.service → /usr/lib/systemd/system/multipathd.service. 2026-03-10T05:32:20.491 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/sockets.target.wants/multipathd.socket → /usr/lib/systemd/system/multipathd.socket. 2026-03-10T05:32:20.491 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:20.676 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 123/146 2026-03-10T05:32:20.699 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 123/146 2026-03-10T05:32:20.699 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:20.699 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T05:32:20.699 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T05:32:20.699 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T05:32:20.699 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:20.755 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 124/146 2026-03-10T05:32:20.758 INFO:teuthology.orchestra.run.vm05.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 124/146 2026-03-10T05:32:20.764 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 125/146 2026-03-10T05:32:20.787 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 126/146 2026-03-10T05:32:20.790 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 127/146 2026-03-10T05:32:20.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:20 vm01 ceph-mon[47941]: pgmap v143: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:20.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:20 vm02 ceph-mon[50473]: pgmap v143: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:21.000 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 135/146 2026-03-10T05:32:21.023 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 135/146 2026-03-10T05:32:21.024 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:21.024 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T05:32:21.024 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T05:32:21.024 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T05:32:21.024 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:21.034 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 136/146 2026-03-10T05:32:21.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:20 vm05 ceph-mon[50927]: pgmap v143: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:21.055 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 136/146 2026-03-10T05:32:21.055 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:21.055 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T05:32:21.056 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:21.213 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 137/146 2026-03-10T05:32:21.234 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 137/146 2026-03-10T05:32:21.234 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:21.234 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T05:32:21.234 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T05:32:21.234 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T05:32:21.234 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:21.322 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 127/146 2026-03-10T05:32:21.327 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 128/146 2026-03-10T05:32:21.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:32:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:32:21.836 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 128/146 2026-03-10T05:32:21.839 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 129/146 2026-03-10T05:32:21.898 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 129/146 2026-03-10T05:32:21.954 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 130/146 2026-03-10T05:32:21.957 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 131/146 2026-03-10T05:32:21.977 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:21 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:21.978 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 131/146 2026-03-10T05:32:21.978 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:21.978 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T05:32:21.978 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T05:32:21.978 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T05:32:21.978 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:21.993 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 132/146 2026-03-10T05:32:22.006 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 132/146 2026-03-10T05:32:22.014 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 133/146 2026-03-10T05:32:22.017 INFO:teuthology.orchestra.run.vm05.stdout: Installing : device-mapper-multipath-0.8.7-45.el9.x86_64 134/146 2026-03-10T05:32:22.030 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: device-mapper-multipath-0.8.7-45.el9.x86_64 134/146 2026-03-10T05:32:22.030 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/sysinit.target.wants/multipathd.service → /usr/lib/systemd/system/multipathd.service. 2026-03-10T05:32:22.030 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/sockets.target.wants/multipathd.socket → /usr/lib/systemd/system/multipathd.socket. 2026-03-10T05:32:22.030 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:22.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:21 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:22.147 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:21 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:22.533 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 135/146 2026-03-10T05:32:22.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:32:22 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:32:22.560 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 135/146 2026-03-10T05:32:22.560 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:22.560 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T05:32:22.560 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T05:32:22.560 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T05:32:22.560 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:22.570 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 136/146 2026-03-10T05:32:22.592 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 136/146 2026-03-10T05:32:22.592 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:22.592 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T05:32:22.592 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:22.752 INFO:teuthology.orchestra.run.vm05.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 137/146 2026-03-10T05:32:22.773 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 137/146 2026-03-10T05:32:22.773 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:32:22.773 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T05:32:22.773 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T05:32:22.773 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T05:32:22.773 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:22 vm05 ceph-mon[50927]: pgmap v144: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:23.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:22 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:22 vm01 ceph-mon[47941]: pgmap v144: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:23.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:22 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:23.147 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:22 vm02 ceph-mon[50473]: pgmap v144: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:23.147 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:22 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:23.724 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 138/146 2026-03-10T05:32:23.736 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 139/146 2026-03-10T05:32:23.742 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 140/146 2026-03-10T05:32:23.798 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 141/146 2026-03-10T05:32:23.808 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 142/146 2026-03-10T05:32:23.812 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 143/146 2026-03-10T05:32:23.812 INFO:teuthology.orchestra.run.vm01.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-10T05:32:23.830 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-10T05:32:23.830 INFO:teuthology.orchestra.run.vm01.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 145/146 2026-03-10T05:32:23.847 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 145/146 2026-03-10T05:32:23.866 INFO:teuthology.orchestra.run.vm01.stdout: Cleanup : kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-10T05:32:25.332 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-10T05:32:25.332 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/146 2026-03-10T05:32:25.332 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/146 2026-03-10T05:32:25.332 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/146 2026-03-10T05:32:25.333 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : device-mapper-multipath-0.8.7-45.el9.x86_64 38/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : device-mapper-multipath-libs-0.8.7-45.el9.x86_64 39/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 40/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 41/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : isns-utils-libs-0.101-4.el9.x86_64 42/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 43/146 2026-03-10T05:32:25.334 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 44/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 45/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 46/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 47/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 48/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 49/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 50/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ply-3.11-14.el9.noarch 51/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 52/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 53/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 54/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 55/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : unzip-6.0-59.el9.x86_64 56/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : zip-3.0-35.el9.x86_64 57/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 58/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 59/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 60/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 61/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 62/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 63/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 64/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 65/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 66/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 67/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 68/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-5.4.4-4.el9.x86_64 69/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 70/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 71/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 72/146 2026-03-10T05:32:25.335 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 73/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 74/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 75/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 76/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 77/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 78/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 79/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 80/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 81/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 138/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 82/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 83/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 84/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 85/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 87/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 88/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 89/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 90/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 91/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 92/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 93/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 94/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 95/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 96/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 97/146 2026-03-10T05:32:25.336 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 98/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 99/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 100/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 101/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 102/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 103/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 104/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 105/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 106/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 107/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 108/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 109/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 110/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 111/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 112/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 113/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 114/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 115/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 116/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 117/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 118/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 119/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 120/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 121/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 122/146 2026-03-10T05:32:25.337 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 123/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 124/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 125/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 126/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 127/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 128/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 129/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 130/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 131/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 132/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 133/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 134/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 135/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 136/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 137/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 138/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : re2-1:20211101-20.el9.x86_64 139/146 2026-03-10T05:32:25.338 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 140/146 2026-03-10T05:32:25.339 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 141/146 2026-03-10T05:32:25.339 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 142/146 2026-03-10T05:32:25.339 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 143/146 2026-03-10T05:32:25.339 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-10T05:32:25.339 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : kpartx-0.8.7-45.el9.x86_64 145/146 2026-03-10T05:32:25.348 INFO:teuthology.orchestra.run.vm05.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 139/146 2026-03-10T05:32:25.353 INFO:teuthology.orchestra.run.vm05.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 140/146 2026-03-10T05:32:25.411 INFO:teuthology.orchestra.run.vm05.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 141/146 2026-03-10T05:32:25.421 INFO:teuthology.orchestra.run.vm05.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 142/146 2026-03-10T05:32:25.424 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 143/146 2026-03-10T05:32:25.425 INFO:teuthology.orchestra.run.vm05.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout:Upgraded: 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: kpartx-0.8.7-45.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout:Installed: 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: device-mapper-multipath-0.8.7-45.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: device-mapper-multipath-libs-0.8.7-45.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3e853.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: isns-utils-libs-0.101-4.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T05:32:25.440 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyparsing-2.4.7-9.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T05:32:25.441 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T05:32:25.442 INFO:teuthology.orchestra.run.vm01.stdout: zip-3.0-35.el9.x86_64 2026-03-10T05:32:25.442 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:25.442 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:32:25.442 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-10T05:32:25.442 INFO:teuthology.orchestra.run.vm05.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 145/146 2026-03-10T05:32:25.460 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 145/146 2026-03-10T05:32:25.479 INFO:teuthology.orchestra.run.vm05.stdout: Cleanup : kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-10T05:32:25.530 DEBUG:teuthology.parallel:result is None 2026-03-10T05:32:25.561 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:25 vm05 ceph-mon[50927]: pgmap v145: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:25.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:25 vm02 ceph-mon[50473]: pgmap v145: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:25.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:25 vm01 ceph-mon[47941]: pgmap v145: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:27.440 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-10T05:32:27.456 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/146 2026-03-10T05:32:27.456 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/146 2026-03-10T05:32:27.456 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/146 2026-03-10T05:32:27.456 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/146 2026-03-10T05:32:27.456 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/146 2026-03-10T05:32:27.456 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/146 2026-03-10T05:32:27.457 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/146 2026-03-10T05:32:27.459 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/146 2026-03-10T05:32:27.459 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/146 2026-03-10T05:32:27.459 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/146 2026-03-10T05:32:27.459 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/146 2026-03-10T05:32:27.459 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : device-mapper-multipath-0.8.7-45.el9.x86_64 38/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : device-mapper-multipath-libs-0.8.7-45.el9.x86_64 39/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9. 40/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3 41/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : isns-utils-libs-0.101-4.el9.x86_64 42/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 43/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 44/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 45/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 46/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 47/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 48/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 49/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 50/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-ply-3.11-14.el9.noarch 51/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 52/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 53/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 54/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 55/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : unzip-6.0-59.el9.x86_64 56/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : zip-3.0-35.el9.x86_64 57/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 58/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 59/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 60/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 61/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 62/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 63/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 64/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 65/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 66/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 67/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 68/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : lua-5.4.4-4.el9.x86_64 69/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 70/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 71/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 72/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 73/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 74/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 75/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 76/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 77/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 78/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 79/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 80/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 81/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 82/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 83/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 84/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 85/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 87/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 88/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 89/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 90/146 2026-03-10T05:32:27.460 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 91/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 92/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 93/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 94/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 95/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 96/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 97/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 98/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 99/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 100/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 101/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 102/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 103/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 104/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 105/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 106/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 107/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 108/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 109/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 110/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 111/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 112/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 113/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 114/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 115/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 116/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 117/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 118/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 119/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 120/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 121/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 122/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 123/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 124/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 125/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 126/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 127/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 128/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 129/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 130/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 131/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 132/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 133/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 134/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 135/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 136/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 137/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 138/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : re2-1:20211101-20.el9.x86_64 139/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 140/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 141/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 142/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 143/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 144/146 2026-03-10T05:32:27.461 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : kpartx-0.8.7-45.el9.x86_64 145/146 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : kpartx-0.8.7-44.el9.x86_64 146/146 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout:Upgraded: 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: kpartx-0.8.7-45.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout:Installed: 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: device-mapper-multipath-0.8.7-45.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: device-mapper-multipath-libs-0.8.7-45.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T05:32:27.734 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: iscsi-initiator-utils-6.2.1.11-0.git4b3e853.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: iscsi-initiator-utils-iscsiuio-6.2.1.11-0.git4b3e853.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: isns-utils-libs-0.101-4.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T05:32:27.735 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyparsing-2.4.7-9.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: zip-3.0-35.el9.x86_64 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:27.736 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:32:27.830 DEBUG:teuthology.parallel:result is None 2026-03-10T05:32:27.830 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:32:27.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:27 vm02 ceph-mon[50473]: pgmap v146: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:28.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:27 vm05 ceph-mon[50927]: pgmap v146: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:28.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:27 vm01 ceph-mon[47941]: pgmap v146: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:28.407 DEBUG:teuthology.orchestra.run.vm01:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-10T05:32:28.428 INFO:teuthology.orchestra.run.vm01.stdout:19.2.3-678.ge911bdeb.el9 2026-03-10T05:32:28.428 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-10T05:32:28.428 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-10T05:32:28.429 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:32:28.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:28 vm02 ceph-mon[50473]: pgmap v147: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:29.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:28 vm05 ceph-mon[50927]: pgmap v147: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:29.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:28 vm01 ceph-mon[47941]: pgmap v147: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:29.101 DEBUG:teuthology.orchestra.run.vm02:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-10T05:32:29.120 INFO:teuthology.orchestra.run.vm02.stdout:19.2.3-678.ge911bdeb.el9 2026-03-10T05:32:29.120 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-10T05:32:29.120 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-10T05:32:29.121 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T05:32:29.753 DEBUG:teuthology.orchestra.run.vm05:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-10T05:32:29.771 INFO:teuthology.orchestra.run.vm05.stdout:19.2.3-678.ge911bdeb.el9 2026-03-10T05:32:29.771 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-10T05:32:29.771 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-10T05:32:29.772 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-10T05:32:29.772 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:32:29.772 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-10T05:32:29.797 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:32:29.797 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-10T05:32:29.822 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T05:32:29.822 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-10T05:32:29.847 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-10T05:32:29.847 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:32:29.847 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/daemon-helper 2026-03-10T05:32:29.872 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-10T05:32:29.937 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:32:29.937 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/daemon-helper 2026-03-10T05:32:29.960 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-10T05:32:30.022 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T05:32:30.022 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/daemon-helper 2026-03-10T05:32:30.046 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-10T05:32:30.109 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-10T05:32:30.109 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:32:30.109 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-10T05:32:30.132 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-10T05:32:30.195 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:32:30.195 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-10T05:32:30.219 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-10T05:32:30.281 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T05:32:30.281 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-10T05:32:30.304 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-10T05:32:30.367 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-10T05:32:30.367 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-10T05:32:30.367 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/stdin-killer 2026-03-10T05:32:30.390 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-10T05:32:30.454 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:32:30.454 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/stdin-killer 2026-03-10T05:32:30.480 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-10T05:32:30.545 DEBUG:teuthology.orchestra.run.vm05:> set -ex 2026-03-10T05:32:30.545 DEBUG:teuthology.orchestra.run.vm05:> sudo dd of=/usr/bin/stdin-killer 2026-03-10T05:32:30.570 DEBUG:teuthology.orchestra.run.vm05:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-10T05:32:30.635 INFO:teuthology.run_tasks:Running task pexec... 2026-03-10T05:32:30.637 INFO:teuthology.task.pexec:Executing custom commands... 2026-03-10T05:32:30.637 DEBUG:teuthology.orchestra.run.vm01:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-10T05:32:30.638 DEBUG:teuthology.orchestra.run.vm02:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-10T05:32:30.638 DEBUG:teuthology.orchestra.run.vm05:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-10T05:32:30.639 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo dnf remove nvme-cli -y 2026-03-10T05:32:30.639 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-10T05:32:30.639 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-10T05:32:30.639 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-10T05:32:30.640 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm01.local 2026-03-10T05:32:30.640 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-10T05:32:30.640 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-10T05:32:30.640 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-10T05:32:30.640 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-10T05:32:30.640 DEBUG:teuthology.task.pexec:ubuntu@vm02.local< sudo dnf remove nvme-cli -y 2026-03-10T05:32:30.640 DEBUG:teuthology.task.pexec:ubuntu@vm02.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-10T05:32:30.640 DEBUG:teuthology.task.pexec:ubuntu@vm02.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-10T05:32:30.640 DEBUG:teuthology.task.pexec:ubuntu@vm02.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-10T05:32:30.640 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm02.local 2026-03-10T05:32:30.640 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-10T05:32:30.640 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-10T05:32:30.640 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-10T05:32:30.640 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-10T05:32:30.676 DEBUG:teuthology.task.pexec:ubuntu@vm05.local< sudo dnf remove nvme-cli -y 2026-03-10T05:32:30.676 DEBUG:teuthology.task.pexec:ubuntu@vm05.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-10T05:32:30.677 DEBUG:teuthology.task.pexec:ubuntu@vm05.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-10T05:32:30.677 DEBUG:teuthology.task.pexec:ubuntu@vm05.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-10T05:32:30.677 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm05.local 2026-03-10T05:32:30.677 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-10T05:32:30.677 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-10T05:32:30.677 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-10T05:32:30.677 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-10T05:32:30.834 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: nvme-cli 2026-03-10T05:32:30.834 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:32:30.835 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: nvme-cli 2026-03-10T05:32:30.835 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:32:30.838 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:32:30.838 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:32:30.838 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:32:30.838 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:32:30.839 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:32:30.839 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:32:30.877 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: nvme-cli 2026-03-10T05:32:30.877 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:32:30.880 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:32:30.880 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:32:30.880 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:32:31.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:30 vm01 ceph-mon[47941]: pgmap v148: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:31.147 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:30 vm02 ceph-mon[50473]: pgmap v148: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:31.209 INFO:teuthology.orchestra.run.vm01.stdout:Last metadata expiration check: 0:00:50 ago on Tue 10 Mar 2026 05:31:41 AM UTC. 2026-03-10T05:32:31.218 INFO:teuthology.orchestra.run.vm02.stdout:Last metadata expiration check: 0:00:52 ago on Tue 10 Mar 2026 05:31:39 AM UTC. 2026-03-10T05:32:31.246 INFO:teuthology.orchestra.run.vm05.stdout:Last metadata expiration check: 0:00:44 ago on Tue 10 Mar 2026 05:31:47 AM UTC. 2026-03-10T05:32:31.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:30 vm05 ceph-mon[50927]: pgmap v148: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:31.309 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:32:31.309 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:32:31.309 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-10T05:32:31.309 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout:Installing: 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout:Installing dependencies: 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout:Install 6 Packages 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout:Total download size: 6.1 M 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout:Installed size: 23 M 2026-03-10T05:32:31.310 INFO:teuthology.orchestra.run.vm01.stdout:Downloading Packages: 2026-03-10T05:32:31.317 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout:Installing: 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout:Installing dependencies: 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout:Install 6 Packages 2026-03-10T05:32:31.318 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:31.319 INFO:teuthology.orchestra.run.vm02.stdout:Total download size: 6.1 M 2026-03-10T05:32:31.319 INFO:teuthology.orchestra.run.vm02.stdout:Installed size: 23 M 2026-03-10T05:32:31.319 INFO:teuthology.orchestra.run.vm02.stdout:Downloading Packages: 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout:Installing: 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout:Installing dependencies: 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout:Install 6 Packages 2026-03-10T05:32:31.344 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:31.345 INFO:teuthology.orchestra.run.vm05.stdout:Total download size: 6.1 M 2026-03-10T05:32:31.345 INFO:teuthology.orchestra.run.vm05.stdout:Installed size: 23 M 2026-03-10T05:32:31.345 INFO:teuthology.orchestra.run.vm05.stdout:Downloading Packages: 2026-03-10T05:32:31.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:32:31 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:32:32.025 INFO:teuthology.orchestra.run.vm02.stdout:(1/6): nvmetcli-0.8-3.el9.noarch.rpm 340 kB/s | 44 kB 00:00 2026-03-10T05:32:32.026 INFO:teuthology.orchestra.run.vm02.stdout:(2/6): python3-configshell-1.1.30-1.el9.noarch. 553 kB/s | 72 kB 00:00 2026-03-10T05:32:32.091 INFO:teuthology.orchestra.run.vm02.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 1.2 MB/s | 84 kB 00:00 2026-03-10T05:32:32.161 INFO:teuthology.orchestra.run.vm02.stdout:(4/6): nvme-cli-2.16-1.el9.x86_64.rpm 4.4 MB/s | 1.2 MB 00:00 2026-03-10T05:32:32.163 INFO:teuthology.orchestra.run.vm02.stdout:(5/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 6.0 MB/s | 837 kB 00:00 2026-03-10T05:32:32.193 INFO:teuthology.orchestra.run.vm05.stdout:(1/6): nvmetcli-0.8-3.el9.noarch.rpm 230 kB/s | 44 kB 00:00 2026-03-10T05:32:32.205 INFO:teuthology.orchestra.run.vm05.stdout:(2/6): python3-configshell-1.1.30-1.el9.noarch. 355 kB/s | 72 kB 00:00 2026-03-10T05:32:32.275 INFO:teuthology.orchestra.run.vm05.stdout:(3/6): nvme-cli-2.16-1.el9.x86_64.rpm 4.2 MB/s | 1.2 MB 00:00 2026-03-10T05:32:32.276 INFO:teuthology.orchestra.run.vm05.stdout:(4/6): python3-kmod-0.9-32.el9.x86_64.rpm 1.0 MB/s | 84 kB 00:00 2026-03-10T05:32:32.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:31 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:32.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:31 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:32.350 INFO:teuthology.orchestra.run.vm05.stdout:(5/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 5.7 MB/s | 837 kB 00:00 2026-03-10T05:32:32.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:31 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:32.659 INFO:teuthology.orchestra.run.vm05.stdout:(6/6): runc-1.4.0-2.el9.x86_64.rpm 10 MB/s | 4.0 MB 00:00 2026-03-10T05:32:32.659 INFO:teuthology.orchestra.run.vm05.stdout:-------------------------------------------------------------------------------- 2026-03-10T05:32:32.659 INFO:teuthology.orchestra.run.vm05.stdout:Total 4.7 MB/s | 6.1 MB 00:01 2026-03-10T05:32:32.714 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-10T05:32:32.721 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-10T05:32:32.722 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-10T05:32:32.790 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-10T05:32:32.790 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-10T05:32:32.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:32:32 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:32:32.960 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-10T05:32:32.976 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-10T05:32:32.983 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 2/6 2026-03-10T05:32:32.990 INFO:teuthology.orchestra.run.vm05.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 3/6 2026-03-10T05:32:32.992 INFO:teuthology.orchestra.run.vm05.stdout: Installing : nvmetcli-0.8-3.el9.noarch 4/6 2026-03-10T05:32:33.038 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 4/6 2026-03-10T05:32:33.165 INFO:teuthology.orchestra.run.vm05.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 5/6 2026-03-10T05:32:33.169 INFO:teuthology.orchestra.run.vm05.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T05:32:33.244 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:32 vm05 ceph-mon[50927]: pgmap v149: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:33.244 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:32 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:33.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:32 vm01 ceph-mon[47941]: pgmap v149: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:33.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:32 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:33.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:32 vm02 ceph-mon[50473]: pgmap v149: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:33.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:32 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:33.513 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T05:32:33.513 INFO:teuthology.orchestra.run.vm05.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T05:32:33.513 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:33.964 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-10T05:32:33.964 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-10T05:32:33.964 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-10T05:32:33.964 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-10T05:32:33.964 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 5/6 2026-03-10T05:32:34.027 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 6/6 2026-03-10T05:32:34.027 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:34.027 INFO:teuthology.orchestra.run.vm05.stdout:Installed: 2026-03-10T05:32:34.027 INFO:teuthology.orchestra.run.vm05.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-10T05:32:34.027 INFO:teuthology.orchestra.run.vm05.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-10T05:32:34.027 INFO:teuthology.orchestra.run.vm05.stdout: python3-urwid-2.1.2-4.el9.x86_64 runc-4:1.4.0-2.el9.x86_64 2026-03-10T05:32:34.027 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:32:34.027 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:32:34.093 DEBUG:teuthology.parallel:result is None 2026-03-10T05:32:35.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:34 vm05 ceph-mon[50927]: pgmap v150: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:35.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:34 vm01 ceph-mon[47941]: pgmap v150: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:35.391 INFO:teuthology.orchestra.run.vm01.stdout:(1/6): nvmetcli-0.8-3.el9.noarch.rpm 347 kB/s | 44 kB 00:00 2026-03-10T05:32:35.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:34 vm02 ceph-mon[50473]: pgmap v150: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:35.438 INFO:teuthology.orchestra.run.vm01.stdout:(2/6): python3-configshell-1.1.30-1.el9.noarch. 414 kB/s | 72 kB 00:00 2026-03-10T05:32:35.474 INFO:teuthology.orchestra.run.vm01.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 1.0 MB/s | 84 kB 00:00 2026-03-10T05:32:35.794 INFO:teuthology.orchestra.run.vm01.stdout:(4/6): runc-1.4.0-2.el9.x86_64.rpm 12 MB/s | 4.0 MB 00:00 2026-03-10T05:32:35.835 INFO:teuthology.orchestra.run.vm01.stdout:(5/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 2.1 MB/s | 837 kB 00:00 2026-03-10T05:32:37.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:36 vm05 ceph-mon[50927]: pgmap v151: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:37.326 INFO:teuthology.orchestra.run.vm02.stdout:(6/6): runc-1.4.0-2.el9.x86_64.rpm 775 kB/s | 4.0 MB 00:05 2026-03-10T05:32:37.326 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------------------------------------------------------- 2026-03-10T05:32:37.326 INFO:teuthology.orchestra.run.vm02.stdout:Total 1.0 MB/s | 6.1 MB 00:06 2026-03-10T05:32:37.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:36 vm01 ceph-mon[47941]: pgmap v151: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:37.379 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-10T05:32:37.390 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-10T05:32:37.390 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-10T05:32:37.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:36 vm02 ceph-mon[50473]: pgmap v151: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:37.458 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-10T05:32:37.459 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-10T05:32:37.625 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-10T05:32:37.639 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-10T05:32:37.645 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 2/6 2026-03-10T05:32:37.652 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 3/6 2026-03-10T05:32:37.653 INFO:teuthology.orchestra.run.vm02.stdout: Installing : nvmetcli-0.8-3.el9.noarch 4/6 2026-03-10T05:32:37.705 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 4/6 2026-03-10T05:32:37.837 INFO:teuthology.orchestra.run.vm02.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 5/6 2026-03-10T05:32:37.841 INFO:teuthology.orchestra.run.vm02.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T05:32:38.195 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T05:32:38.195 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T05:32:38.195 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:38.654 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-10T05:32:38.654 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-10T05:32:38.654 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-10T05:32:38.654 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-10T05:32:38.654 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 5/6 2026-03-10T05:32:38.728 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 6/6 2026-03-10T05:32:38.728 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:38.728 INFO:teuthology.orchestra.run.vm02.stdout:Installed: 2026-03-10T05:32:38.728 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-10T05:32:38.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-10T05:32:38.728 INFO:teuthology.orchestra.run.vm02.stdout: python3-urwid-2.1.2-4.el9.x86_64 runc-4:1.4.0-2.el9.x86_64 2026-03-10T05:32:38.728 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:32:38.728 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:32:38.790 DEBUG:teuthology.parallel:result is None 2026-03-10T05:32:39.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:38 vm05 ceph-mon[50927]: pgmap v152: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:39.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:38 vm01 ceph-mon[47941]: pgmap v152: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:39.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:38 vm02 ceph-mon[50473]: pgmap v152: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:40.168 INFO:teuthology.orchestra.run.vm01.stdout:(6/6): nvme-cli-2.16-1.el9.x86_64.rpm 240 kB/s | 1.2 MB 00:04 2026-03-10T05:32:40.168 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-10T05:32:40.168 INFO:teuthology.orchestra.run.vm01.stdout:Total 708 kB/s | 6.1 MB 00:08 2026-03-10T05:32:40.229 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-10T05:32:40.237 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-10T05:32:40.237 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-10T05:32:40.308 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-10T05:32:40.308 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-10T05:32:40.483 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-10T05:32:40.498 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-10T05:32:40.505 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 2/6 2026-03-10T05:32:40.512 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 3/6 2026-03-10T05:32:40.514 INFO:teuthology.orchestra.run.vm01.stdout: Installing : nvmetcli-0.8-3.el9.noarch 4/6 2026-03-10T05:32:40.558 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 4/6 2026-03-10T05:32:40.700 INFO:teuthology.orchestra.run.vm01.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 5/6 2026-03-10T05:32:40.704 INFO:teuthology.orchestra.run.vm01.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T05:32:41.043 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T05:32:41.043 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T05:32:41.043 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:41.052 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:40 vm01 ceph-mon[47941]: pgmap v153: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:41.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:40 vm05 ceph-mon[50927]: pgmap v153: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:41.339 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:32:41 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:32:41.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:40 vm02 ceph-mon[50473]: pgmap v153: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:41.496 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-10T05:32:41.496 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-10T05:32:41.496 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-10T05:32:41.496 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-10T05:32:41.496 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 5/6 2026-03-10T05:32:41.558 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 6/6 2026-03-10T05:32:41.558 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:41.558 INFO:teuthology.orchestra.run.vm01.stdout:Installed: 2026-03-10T05:32:41.558 INFO:teuthology.orchestra.run.vm01.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-10T05:32:41.558 INFO:teuthology.orchestra.run.vm01.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-10T05:32:41.558 INFO:teuthology.orchestra.run.vm01.stdout: python3-urwid-2.1.2-4.el9.x86_64 runc-4:1.4.0-2.el9.x86_64 2026-03-10T05:32:41.558 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:32:41.558 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:32:41.617 DEBUG:teuthology.parallel:result is None 2026-03-10T05:32:41.617 INFO:teuthology.run_tasks:Running task ceph_iscsi_client... 2026-03-10T05:32:41.620 INFO:tasks.ceph_iscsi_client:Setting up ceph-iscsi client... 2026-03-10T05:32:41.620 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:32:41.620 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /etc/iscsi 2026-03-10T05:32:41.620 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/iscsi/initiatorname.iscsi 2026-03-10T05:32:41.657 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl restart iscsid 2026-03-10T05:32:41.756 DEBUG:teuthology.orchestra.run.vm02:> sudo modprobe dm_multipath 2026-03-10T05:32:41.786 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-10T05:32:41.786 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/multipath.conf 2026-03-10T05:32:41.849 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl start multipathd 2026-03-10T05:32:41.955 INFO:teuthology.run_tasks:Running task cram... 2026-03-10T05:32:41.959 INFO:tasks.cram:Pulling tests from https://github.com/kshtsk/ceph.git ref 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b 2026-03-10T05:32:41.959 DEBUG:teuthology.orchestra.run.vm01:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.0 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-10T05:32:42.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:41 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:42.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:41 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:42.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:41 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:42.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:32:42 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:32:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:42 vm05 ceph-mon[50927]: pgmap v154: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:42 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:43.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:42 vm01 ceph-mon[47941]: pgmap v154: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:43.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:42 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:43.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:42 vm02 ceph-mon[50473]: pgmap v154: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:43.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:42 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:43.469 INFO:teuthology.orchestra.run.vm01.stdout:Collecting cram==0.6 2026-03-10T05:32:43.499 INFO:teuthology.orchestra.run.vm01.stdout: Downloading cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-10T05:32:43.563 INFO:teuthology.orchestra.run.vm01.stdout:Installing collected packages: cram 2026-03-10T05:32:43.631 INFO:teuthology.orchestra.run.vm01.stdout:Successfully installed cram-0.6 2026-03-10T05:32:43.729 INFO:teuthology.orchestra.run.vm01.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-03-10T05:32:43.729 INFO:teuthology.orchestra.run.vm01.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/virtualenv/bin/python3 -m pip install --upgrade pip' command. 2026-03-10T05:32:43.759 DEBUG:teuthology.orchestra.run.vm01:> rm -rf /home/ubuntu/cephtest/clone.client.0 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.0 && cd /home/ubuntu/cephtest/clone.client.0 && git checkout 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b 2026-03-10T05:32:43.775 INFO:teuthology.orchestra.run.vm01.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.0'... 2026-03-10T05:32:45.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:44 vm05 ceph-mon[50927]: pgmap v155: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:45.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:44 vm01 ceph-mon[47941]: pgmap v155: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:45.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:44 vm02 ceph-mon[50473]: pgmap v155: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:47.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:46 vm05 ceph-mon[50927]: pgmap v156: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:47.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:46 vm01 ceph-mon[47941]: pgmap v156: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:47.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:46 vm02 ceph-mon[50473]: pgmap v156: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:49.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:48 vm05 ceph-mon[50927]: pgmap v157: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:49.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:48 vm01 ceph-mon[47941]: pgmap v157: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:49.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:48 vm02 ceph-mon[50473]: pgmap v157: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:51.301 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:51 vm01 ceph-mon[47941]: pgmap v158: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:51.301 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:32:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:32:51.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:51 vm05 ceph-mon[50927]: pgmap v158: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:51.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:51 vm02 ceph-mon[50473]: pgmap v158: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:52.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:52 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:52.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:52 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:52.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:52 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:52.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:32:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:32:53.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:53 vm05 ceph-mon[50927]: pgmap v159: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:53.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:53 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:53.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:53 vm01 ceph-mon[47941]: pgmap v159: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:53.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:53 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:53 vm02 ceph-mon[50473]: pgmap v159: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:53 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:32:55.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:55 vm05 ceph-mon[50927]: pgmap v160: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:55 vm01 ceph-mon[47941]: pgmap v160: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:55.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:55 vm02 ceph-mon[50473]: pgmap v160: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:57.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:57 vm05 ceph-mon[50927]: pgmap v161: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:57.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:57 vm01 ceph-mon[47941]: pgmap v161: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:57.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:57 vm02 ceph-mon[50473]: pgmap v161: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:32:59.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:32:59 vm05 ceph-mon[50927]: pgmap v162: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:59.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:32:59 vm01 ceph-mon[47941]: pgmap v162: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:32:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:32:59 vm02 ceph-mon[50473]: pgmap v162: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:01.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:01 vm01 ceph-mon[47941]: pgmap v163: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:01.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:33:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:33:01.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:01 vm02 ceph-mon[50473]: pgmap v163: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:01.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:01 vm05 ceph-mon[50927]: pgmap v163: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:02.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:02 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:02.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:02 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:02.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:33:02 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:33:02.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:02 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:03 vm01 ceph-mon[47941]: pgmap v164: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:03 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:03.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:03 vm02 ceph-mon[50473]: pgmap v164: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:03.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:03 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:03 vm05 ceph-mon[50927]: pgmap v164: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:03 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:05.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:05 vm01 ceph-mon[47941]: pgmap v165: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:05.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:05 vm02 ceph-mon[50473]: pgmap v165: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:05.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:05 vm05 ceph-mon[50927]: pgmap v165: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:08.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:08 vm05 ceph-mon[50927]: pgmap v166: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:08.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:08 vm01 ceph-mon[47941]: pgmap v166: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:08.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:08 vm02 ceph-mon[50473]: pgmap v166: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:09.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:09 vm02 ceph-mon[50473]: pgmap v167: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:09.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:09 vm05 ceph-mon[50927]: pgmap v167: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:09.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:09 vm01 ceph-mon[47941]: pgmap v167: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:11.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:11 vm02 ceph-mon[50473]: pgmap v168: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:11.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:11 vm05 ceph-mon[50927]: pgmap v168: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:11.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:11 vm01 ceph-mon[47941]: pgmap v168: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:11.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:33:11 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:33:12.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:12 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:12.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:33:12 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:33:12.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:12 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:12.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:12 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:13.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:13 vm02 ceph-mon[50473]: pgmap v169: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:13 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:13.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:13 vm05 ceph-mon[50927]: pgmap v169: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:13.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:13 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:13.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:13 vm01 ceph-mon[47941]: pgmap v169: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:13.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:13 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:14 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:33:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:14 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:33:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:14 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:33:14.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:14 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:33:14.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:14 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:33:14.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:14 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:33:14.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:14 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:33:14.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:14 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:33:14.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:33:14.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:33:14.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:33:14.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:33:15.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:15 vm02 ceph-mon[50473]: pgmap v170: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:15.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:15 vm05 ceph-mon[50927]: pgmap v170: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:15.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:15 vm01 ceph-mon[47941]: pgmap v170: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:17.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:17 vm05 ceph-mon[50927]: pgmap v171: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:17.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:17 vm01 ceph-mon[47941]: pgmap v171: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:17.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:17 vm02 ceph-mon[50473]: pgmap v171: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:19.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:19 vm05 ceph-mon[50927]: pgmap v172: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:19.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:19 vm01 ceph-mon[47941]: pgmap v172: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:19.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:19 vm02 ceph-mon[50473]: pgmap v172: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:21.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:21 vm05 ceph-mon[50927]: pgmap v173: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:21.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:21 vm01 ceph-mon[47941]: pgmap v173: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:21.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:33:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:33:21.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:21 vm02 ceph-mon[50473]: pgmap v173: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:22.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:33:22 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:33:22.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:22 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:22.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:22 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:22.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:22 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:23.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:23 vm05 ceph-mon[50927]: pgmap v174: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:23.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:23 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:23.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:23 vm01 ceph-mon[47941]: pgmap v174: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:23.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:23 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:23.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:23 vm02 ceph-mon[50473]: pgmap v174: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:23.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:23 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr:Note: switching to '75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b'. 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr:state without impacting any branches by switching back to a branch. 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr: git switch -c 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr:Or undo this operation with: 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr: git switch - 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:33:24.589 INFO:teuthology.orchestra.run.vm01.stderr:HEAD is now at 75a68fd8ca3 qa/suites/orch/cephadm/osds: drop nvme_loop task 2026-03-10T05:33:24.595 DEBUG:teuthology.orchestra.run.vm01:> cp -- /home/ubuntu/cephtest/clone.client.0/src/test/cli-integration/rbd/gwcli_create.t /home/ubuntu/cephtest/archive/cram.client.0 2026-03-10T05:33:24.651 DEBUG:teuthology.orchestra.run.vm02:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.1 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-10T05:33:25.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:25 vm05 ceph-mon[50927]: pgmap v175: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:25.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:25 vm01 ceph-mon[47941]: pgmap v175: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:25.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:25 vm02 ceph-mon[50473]: pgmap v175: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:26.067 INFO:teuthology.orchestra.run.vm02.stdout:Collecting cram==0.6 2026-03-10T05:33:26.097 INFO:teuthology.orchestra.run.vm02.stdout: Downloading cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-10T05:33:26.108 INFO:teuthology.orchestra.run.vm02.stdout:Installing collected packages: cram 2026-03-10T05:33:26.115 INFO:teuthology.orchestra.run.vm02.stdout:Successfully installed cram-0.6 2026-03-10T05:33:26.196 INFO:teuthology.orchestra.run.vm02.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-03-10T05:33:26.196 INFO:teuthology.orchestra.run.vm02.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/virtualenv/bin/python3 -m pip install --upgrade pip' command. 2026-03-10T05:33:26.226 DEBUG:teuthology.orchestra.run.vm02:> rm -rf /home/ubuntu/cephtest/clone.client.1 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.1 && cd /home/ubuntu/cephtest/clone.client.1 && git checkout 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b 2026-03-10T05:33:26.245 INFO:teuthology.orchestra.run.vm02.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.1'... 2026-03-10T05:33:27.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:27 vm05 ceph-mon[50927]: pgmap v176: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:27.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:27 vm01 ceph-mon[47941]: pgmap v176: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:27.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:27 vm02 ceph-mon[50473]: pgmap v176: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:29.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:29 vm05 ceph-mon[50927]: pgmap v177: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:29.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:29 vm01 ceph-mon[47941]: pgmap v177: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:29.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:29 vm02 ceph-mon[50473]: pgmap v177: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:31.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:31 vm05 ceph-mon[50927]: pgmap v178: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:31.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:33:31 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:33:31.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:31 vm01 ceph-mon[47941]: pgmap v178: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:31.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:31 vm02 ceph-mon[50473]: pgmap v178: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:32.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:33:32 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:33:32.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:32 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:32.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:32 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:32.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:32 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:33.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:33 vm05 ceph-mon[50927]: pgmap v179: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:33.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:33 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:33.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:33 vm01 ceph-mon[47941]: pgmap v179: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:33.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:33 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:33.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:33 vm02 ceph-mon[50473]: pgmap v179: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:33.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:33 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:35.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:35 vm05 ceph-mon[50927]: pgmap v180: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:35.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:35 vm01 ceph-mon[47941]: pgmap v180: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:35.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:35 vm02 ceph-mon[50473]: pgmap v180: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:37.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:37 vm05 ceph-mon[50927]: pgmap v181: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:37.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:37 vm01 ceph-mon[47941]: pgmap v181: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:37.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:37 vm02 ceph-mon[50473]: pgmap v181: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:39.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:39 vm01 ceph-mon[47941]: pgmap v182: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:39.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:39 vm02 ceph-mon[50473]: pgmap v182: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:39.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:39 vm05 ceph-mon[50927]: pgmap v182: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:41.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:40 vm05 ceph-mon[50927]: pgmap v183: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:41.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:40 vm01 ceph-mon[47941]: pgmap v183: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:41.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:40 vm02 ceph-mon[50473]: pgmap v183: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:41.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:33:41 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:33:42.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:41 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:42.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:41 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:42.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:41 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:42.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:33:42 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:33:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:43 vm05 ceph-mon[50927]: pgmap v184: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:43 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:43.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:42 vm01 ceph-mon[47941]: pgmap v184: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:43.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:43 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:43.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:42 vm02 ceph-mon[50473]: pgmap v184: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:42 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:45.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:45 vm05 ceph-mon[50927]: pgmap v185: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:45.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:45 vm01 ceph-mon[47941]: pgmap v185: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:45.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:45 vm02 ceph-mon[50473]: pgmap v185: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:47.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:47 vm05 ceph-mon[50927]: pgmap v186: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:47.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:47 vm01 ceph-mon[47941]: pgmap v186: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:47.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:47 vm02 ceph-mon[50473]: pgmap v186: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:49.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:49 vm05 ceph-mon[50927]: pgmap v187: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:49.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:49 vm01 ceph-mon[47941]: pgmap v187: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:49 vm02 ceph-mon[50473]: pgmap v187: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:51 vm05 ceph-mon[50927]: pgmap v188: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:51.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:51 vm01 ceph-mon[47941]: pgmap v188: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:51.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:33:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:33:51.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:51 vm02 ceph-mon[50473]: pgmap v188: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:52.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:33:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:33:52.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:52 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:52.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:52 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:52.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:52 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:53 vm05 ceph-mon[50927]: pgmap v189: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:53 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:53.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:53 vm01 ceph-mon[47941]: pgmap v189: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:53.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:53 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:53 vm02 ceph-mon[50473]: pgmap v189: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:53 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:33:55.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:55 vm05 ceph-mon[50927]: pgmap v190: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:55.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:55 vm01 ceph-mon[47941]: pgmap v190: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:55.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:55 vm02 ceph-mon[50473]: pgmap v190: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:57.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:57 vm01 ceph-mon[47941]: pgmap v191: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:57.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:57 vm02 ceph-mon[50473]: pgmap v191: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:57.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:57 vm05 ceph-mon[50927]: pgmap v191: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:33:59.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:33:59 vm01 ceph-mon[47941]: pgmap v192: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:33:59 vm02 ceph-mon[50473]: pgmap v192: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:33:59.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:33:59 vm05 ceph-mon[50927]: pgmap v192: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:01.587 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:34:01.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:01 vm01 ceph-mon[47941]: pgmap v193: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:01.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:01 vm02 ceph-mon[50473]: pgmap v193: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:01.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:01 vm05 ceph-mon[50927]: pgmap v193: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:02.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:02 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:02.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:02 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:02.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:02 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:34:02.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:02 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:03.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:03 vm01 ceph-mon[47941]: pgmap v194: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:03.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:03 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:03.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:03 vm02 ceph-mon[50473]: pgmap v194: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:03.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:03 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:03 vm05 ceph-mon[50927]: pgmap v194: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:03 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:05.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:05 vm02 ceph-mon[50473]: pgmap v195: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:05.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:05 vm05 ceph-mon[50927]: pgmap v195: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:05.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:05 vm01 ceph-mon[47941]: pgmap v195: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:07.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:07 vm02 ceph-mon[50473]: pgmap v196: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr:Note: switching to '75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b'. 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr:state without impacting any branches by switching back to a branch. 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr: git switch -c 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr:Or undo this operation with: 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr: git switch - 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:34:07.727 INFO:teuthology.orchestra.run.vm02.stderr:HEAD is now at 75a68fd8ca3 qa/suites/orch/cephadm/osds: drop nvme_loop task 2026-03-10T05:34:07.732 DEBUG:teuthology.orchestra.run.vm02:> cp -- /home/ubuntu/cephtest/clone.client.1/src/test/cli-integration/rbd/iscsi_client.t /home/ubuntu/cephtest/archive/cram.client.1 2026-03-10T05:34:07.789 DEBUG:teuthology.orchestra.run.vm05:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.2 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-10T05:34:07.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:07 vm05 ceph-mon[50927]: pgmap v196: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:07.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:07 vm01 ceph-mon[47941]: pgmap v196: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:09.280 INFO:teuthology.orchestra.run.vm05.stdout:Collecting cram==0.6 2026-03-10T05:34:09.317 INFO:teuthology.orchestra.run.vm05.stdout: Downloading cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-10T05:34:09.329 INFO:teuthology.orchestra.run.vm05.stdout:Installing collected packages: cram 2026-03-10T05:34:09.336 INFO:teuthology.orchestra.run.vm05.stdout:Successfully installed cram-0.6 2026-03-10T05:34:09.422 INFO:teuthology.orchestra.run.vm05.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-03-10T05:34:09.422 INFO:teuthology.orchestra.run.vm05.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/virtualenv/bin/python3 -m pip install --upgrade pip' command. 2026-03-10T05:34:09.465 DEBUG:teuthology.orchestra.run.vm05:> rm -rf /home/ubuntu/cephtest/clone.client.2 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.2 && cd /home/ubuntu/cephtest/clone.client.2 && git checkout 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b 2026-03-10T05:34:09.522 INFO:teuthology.orchestra.run.vm05.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.2'... 2026-03-10T05:34:09.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:09 vm02 ceph-mon[50473]: pgmap v197: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:09.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:09 vm05 ceph-mon[50927]: pgmap v197: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:09.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:09 vm01 ceph-mon[47941]: pgmap v197: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:11.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:11 vm05 ceph-mon[50927]: pgmap v198: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:11.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:11 vm01 ceph-mon[47941]: pgmap v198: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:11.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:11 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:34:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:11 vm02 ceph-mon[50473]: pgmap v198: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:12.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:12 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:12.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:12 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:34:12.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:12 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:12.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:12 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:13.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:13 vm05 ceph-mon[50927]: pgmap v199: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:13.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:13 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:13.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:13 vm01 ceph-mon[47941]: pgmap v199: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:13.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:13 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:13.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:13 vm02 ceph-mon[50473]: pgmap v199: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:13.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:13 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:14.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:14 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:34:14.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:14 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:34:14.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:14 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:34:14.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:14 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:34:14.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:34:14.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:34:14.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:34:14.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:14 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:34:14.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:14 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:34:14.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:14 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:34:14.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:14 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:34:14.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:14 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:34:15.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:15 vm05 ceph-mon[50927]: pgmap v200: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:15.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:15 vm01 ceph-mon[47941]: pgmap v200: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:15.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:15 vm02 ceph-mon[50473]: pgmap v200: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:17.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:17 vm05 ceph-mon[50927]: pgmap v201: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:17.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:17 vm01 ceph-mon[47941]: pgmap v201: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:17.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:17 vm02 ceph-mon[50473]: pgmap v201: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:19.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:19 vm05 ceph-mon[50927]: pgmap v202: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:19.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:19 vm01 ceph-mon[47941]: pgmap v202: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:19.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:19 vm02 ceph-mon[50473]: pgmap v202: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:21.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:34:21.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:21 vm01 ceph-mon[47941]: pgmap v203: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:21.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:21 vm02 ceph-mon[50473]: pgmap v203: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:22.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:21 vm05 ceph-mon[50927]: pgmap v203: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:22.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:22 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:22.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:22 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:34:22.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:22 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:22.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:22 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:24.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:24 vm01 ceph-mon[47941]: pgmap v204: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:24.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:24 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:24.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:24 vm02 ceph-mon[50473]: pgmap v204: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:24.407 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:24 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:24.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:24 vm05 ceph-mon[50927]: pgmap v204: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:24.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:24 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:25.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:25 vm05 ceph-mon[50927]: pgmap v205: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:25.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:25 vm01 ceph-mon[47941]: pgmap v205: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:25.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:25 vm02 ceph-mon[50473]: pgmap v205: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:27.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:27 vm05 ceph-mon[50927]: pgmap v206: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:27.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:27 vm01 ceph-mon[47941]: pgmap v206: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:27.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:27 vm02 ceph-mon[50473]: pgmap v206: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:29.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:29 vm05 ceph-mon[50927]: pgmap v207: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:29.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:29 vm01 ceph-mon[47941]: pgmap v207: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:29.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:29 vm02 ceph-mon[50473]: pgmap v207: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:31.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:31 vm05 ceph-mon[50927]: pgmap v208: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:31.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:31 vm01 ceph-mon[47941]: pgmap v208: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:31.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:31 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:34:31.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:31 vm02 ceph-mon[50473]: pgmap v208: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:32.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:32 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:32.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:32 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:34:32.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:32 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:32.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:32 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:33.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:33 vm05 ceph-mon[50927]: pgmap v209: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:33.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:33 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:33.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:33 vm01 ceph-mon[47941]: pgmap v209: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:33.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:33 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:33.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:33 vm02 ceph-mon[50473]: pgmap v209: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:33.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:33 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:35.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:35 vm05 ceph-mon[50927]: pgmap v210: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:35.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:35 vm01 ceph-mon[47941]: pgmap v210: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:35.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:35 vm02 ceph-mon[50473]: pgmap v210: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:38.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:37 vm05 ceph-mon[50927]: pgmap v211: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:38.087 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:37 vm01 ceph-mon[47941]: pgmap v211: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:38.147 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:37 vm02 ceph-mon[50473]: pgmap v211: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:39.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:38 vm05 ceph-mon[50927]: pgmap v212: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:39.087 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:38 vm01 ceph-mon[47941]: pgmap v212: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:39.147 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:38 vm02 ceph-mon[50473]: pgmap v212: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:41.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:40 vm05 ceph-mon[50927]: pgmap v213: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:41.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:40 vm01 ceph-mon[47941]: pgmap v213: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:41.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:40 vm02 ceph-mon[50473]: pgmap v213: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:41.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:41 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:34:42.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:41 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:42.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:41 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:42.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:41 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:42.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:42 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:34:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:42 vm05 ceph-mon[50927]: pgmap v214: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:42 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:43.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:42 vm01 ceph-mon[47941]: pgmap v214: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:43.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:42 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:42 vm02 ceph-mon[50473]: pgmap v214: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:42 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:45.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:45 vm05 ceph-mon[50927]: pgmap v215: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:45.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:45 vm01 ceph-mon[47941]: pgmap v215: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:45.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:45 vm02 ceph-mon[50473]: pgmap v215: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:47.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:47 vm01 ceph-mon[47941]: pgmap v216: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:47.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:47 vm02 ceph-mon[50473]: pgmap v216: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:47.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:47 vm05 ceph-mon[50927]: pgmap v216: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:49.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:49 vm02 ceph-mon[50473]: pgmap v217: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr:Note: switching to '75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b'. 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr:state without impacting any branches by switching back to a branch. 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr: git switch -c 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr:Or undo this operation with: 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr: git switch - 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:34:49.449 INFO:teuthology.orchestra.run.vm05.stderr:HEAD is now at 75a68fd8ca3 qa/suites/orch/cephadm/osds: drop nvme_loop task 2026-03-10T05:34:49.454 DEBUG:teuthology.orchestra.run.vm05:> cp -- /home/ubuntu/cephtest/clone.client.2/src/test/cli-integration/rbd/gwcli_delete.t /home/ubuntu/cephtest/archive/cram.client.2 2026-03-10T05:34:49.511 INFO:tasks.cram:Running tests for client.0... 2026-03-10T05:34:49.511 DEBUG:teuthology.orchestra.run.vm01:> CEPH_REF=master CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.0/*.t 2026-03-10T05:34:49.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:49 vm05 ceph-mon[50927]: pgmap v217: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:49.555 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:49 vm01 ceph-mon[47941]: pgmap v217: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:34:50.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:49 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:49] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:50.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:49 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:49] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:50.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3216880099' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:50.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1124212593' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:50.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4222371711' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:50.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1220815257' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:50.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2714870094' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:50.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:50 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3216880099' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:50.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:50 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1124212593' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:50.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:50 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4222371711' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:50.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:50 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1220815257' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:50.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:50 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2714870094' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:50.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:50 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3216880099' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:50.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:50 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1124212593' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:50.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:50 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4222371711' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:50.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:50 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1220815257' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:50.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:50 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2714870094' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:50.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:50] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:50] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (LUN.allocate) created datapool/block0 successfully 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block0' to LIO backstore user:rbd 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: /usr/lib/python3.9/site-packages/rtslib_fb/root.py:180: UserWarning: Cannot set dbroot to /var/target. Target devices have already been registered. 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: warn("Cannot set dbroot to {}. Target devices have already been registered." 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (LUN.add_dev_to_lio) Successfully added datapool/block0 to LIO 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (LUN.allocate) added 'datapool/block0' to LIO and config object 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:50] "PUT /api/_disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:50] "PUT /api/_disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _disk update on localhost, successful 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:50] "PUT /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:50] "PUT /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:50] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:50] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:50] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:50] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:50] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:50] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:50] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:50.638 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:50] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:51.055 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:50 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug Change detected - internal 0 / xattr 1 refreshing 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: pgmap v218: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4117069054' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2531138562' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/451014141' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3274198358' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1732746430' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/81429086' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/81429086' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/81429086' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1235753545' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2476942332' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4240982538' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2618652738' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1907720217' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:51.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4168435081' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:51.339 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:50] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:51.339 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:50 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:50] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:51.339 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:51] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:51.339 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:51] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:51.339 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:51] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:51.339 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:51] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: pgmap v218: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4117069054' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2531138562' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/451014141' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3274198358' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1732746430' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/81429086' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/81429086' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/81429086' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1235753545' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2476942332' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4240982538' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2618652738' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1907720217' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:51 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4168435081' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: pgmap v218: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4117069054' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2531138562' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/451014141' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3274198358' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1732746430' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/81429086' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/81429086' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/81429086' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1235753545' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2476942332' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4240982538' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2618652738' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1907720217' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:51.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:51 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4168435081' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:51.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:34:51.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:51] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:51.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:51] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:51.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:51] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:51.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:51] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:51.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:51] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:51.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:51] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:51] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:51] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:52] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:52] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:52] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:52.339 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:52] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:52.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:52.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/328755379' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:52.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/787558137' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:52.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/717509329' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:52.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1974278525' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:52.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2906820489' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:52.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/200191385' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:52.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1921528379' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:52.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/850222063' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:52.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3619551169' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:52.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/450151237' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:52 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/328755379' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/787558137' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/717509329' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1974278525' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2906820489' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/200191385' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1921528379' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/850222063' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3619551169' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/450151237' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:52.441 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:52 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:52.441 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/328755379' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:52.441 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/787558137' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:52.441 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/717509329' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:52.441 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1974278525' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:52.441 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2906820489' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:52.441 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/200191385' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:52.441 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1921528379' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:52.441 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/850222063' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:52.441 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3619551169' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:52.441 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/450151237' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:52.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:34:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:52] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:52] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:52] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:52] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Gateway.create_target) created an iscsi target with iqn of 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-10T05:34:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:52] "PUT /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:34:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:52] "PUT /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:34:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:52] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:52] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:52] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:52] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:53.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:52] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:53.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:52] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:53.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:53] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:53.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:53] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:53.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:53] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:53.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:53] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:53.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:53] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:53.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:53] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:53.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-mon[47941]: pgmap v219: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.8 KiB/s rd, 85 B/s wr, 2 op/s 2026-03-10T05:34:53.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3840103216' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:53.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3551649845' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:53.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2938070281' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:53.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:53.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/260269840' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:53.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/168335116' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:53.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4191705202' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:53.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4136470858' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:53.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3672256145' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:53.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1065301026' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:53.339 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/80080212' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:53 vm02 ceph-mon[50473]: pgmap v219: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.8 KiB/s rd, 85 B/s wr, 2 op/s 2026-03-10T05:34:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:53 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3840103216' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:53 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3551649845' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:53 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2938070281' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:53 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:53 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/260269840' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:53 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/168335116' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:53 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4191705202' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:53 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4136470858' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:53 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3672256145' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:53 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1065301026' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:53 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/80080212' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:53 vm05 ceph-mon[50927]: pgmap v219: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.8 KiB/s rd, 85 B/s wr, 2 op/s 2026-03-10T05:34:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:53 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3840103216' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:53 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3551649845' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:53 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2938070281' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:53 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:34:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:53 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/260269840' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:53 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/168335116' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:53 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4191705202' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:53 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4136470858' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:53 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3672256145' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:53 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1065301026' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:53.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:53 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/80080212' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:53.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:53] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:53.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:53] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:53.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:53] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:53.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:53] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:53.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:53] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:53.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:53] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:53.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:53] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:53.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:53] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:54.055 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:53 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug Change detected - internal 1 / xattr 2 refreshing 2026-03-10T05:34:54.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:53] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:54.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:53 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:53] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:54.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:54] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:54.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:54] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:54.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:54.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:54.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:54.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:54.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1957865471' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:54.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2646223855' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:54.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3906331647' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:54.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2031479346' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:54.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/705933722' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:54.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1593275618' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:54.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2682589862' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:54.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/324519963' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:54.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2513284946' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:54.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2198343373' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:54.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:54 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1957865471' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:54.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:54 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2646223855' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:54.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:54 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3906331647' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:54.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:54 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2031479346' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:54.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:54 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/705933722' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:54.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:54 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1593275618' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:54.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:54 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2682589862' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:54.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:54 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/324519963' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:54.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:54 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2513284946' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:54.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:54 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2198343373' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:54.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:54 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1957865471' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:54.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:54 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2646223855' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:54.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:54 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3906331647' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:54.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:54 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2031479346' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:54.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:54 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/705933722' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:54.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:54 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1593275618' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:54.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:54 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2682589862' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:54.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:54 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/324519963' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:54.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:54 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2513284946' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:54.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:54 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2198343373' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:54.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:54.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:54.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:54] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:54.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:54] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:54.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:54.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:54.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:54.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:55.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:55.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:55.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:54] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:55.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:54] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:55.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:55.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:55.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:55.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:54 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:54] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-mon[47941]: pgmap v220: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.8 KiB/s rd, 85 B/s wr, 2 op/s 2026-03-10T05:34:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3195187102' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2677606995' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2267240946' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1339627440' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1263375624' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2516596524' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/813408973' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3160014809' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/81025221' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:55.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2199657465' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:55.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:55 vm02 ceph-mon[50473]: pgmap v220: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.8 KiB/s rd, 85 B/s wr, 2 op/s 2026-03-10T05:34:55.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:55 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3195187102' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:55.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:55 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2677606995' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:55.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:55 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2267240946' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:55.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:55 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1339627440' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:55.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:55 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1263375624' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:55.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:55 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2516596524' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:55.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:55 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/813408973' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:55.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:55 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3160014809' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:55.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:55 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/81025221' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:55.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:55 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2199657465' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:55.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:55 vm05 ceph-mon[50927]: pgmap v220: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.8 KiB/s rd, 85 B/s wr, 2 op/s 2026-03-10T05:34:55.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:55 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3195187102' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:55.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:55 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2677606995' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:55.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:55 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2267240946' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:55.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:55 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1339627440' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:55.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:55 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1263375624' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:55.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:55 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2516596524' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:55.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:55 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/813408973' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:55.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:55 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3160014809' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:55.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:55 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/81025221' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:55.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:55 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2199657465' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:55.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:55] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:55.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:55] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:55.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:55] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:55.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:55] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:55.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:55] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:55.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:55] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:55.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:55] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:55.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:55] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.142 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:55] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.142 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:55] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.142 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:55] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:56.142 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:55] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:56.142 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:55] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.142 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:55] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.142 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:55] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:55] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:55] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:55] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:55] "GET /api/sysinfo/hostname HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:55] "GET /api/sysinfo/hostname HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug gateway validation needed for vm01.local 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:55] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:55] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:55] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:55] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:55] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:55] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:55] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:55] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:55 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug An additional 1 tpg's are required 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Gateway.create_tpg) created TPG '1' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug iqn.2003-01.com.redhat.iscsi-gw:ceph-gw - Adding the IP to the enabled tpg, allowing iSCSI logins 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:56] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:56] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _gateway update on localhost, successful 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:56] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:56] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:56] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:56.143 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:56] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:56.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:56 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1790523428' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:56.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:56 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/392408637' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:56.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:56 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1922002800' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:56.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:56 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1480056228' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:56.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:56 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3578365079' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:56.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:56 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2254354284' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:56.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:56 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2507285303' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:56.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:56 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2597916702' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:56.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:56 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2035421076' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:56.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:56 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/958667374' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:56.411 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1790523428' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:56.411 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/392408637' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:56.411 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1922002800' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:56.411 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1480056228' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:56.411 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3578365079' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:56.411 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2254354284' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:56.411 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2507285303' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:56.411 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2597916702' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:56.411 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2035421076' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:56.411 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/958667374' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:56.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:56 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1790523428' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:56.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:56 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/392408637' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:56.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:56 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1922002800' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:56.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:56 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1480056228' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:56.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:56 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3578365079' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:56.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:56 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2254354284' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:56.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:56 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2507285303' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:56.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:56 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2597916702' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:56.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:56 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2035421076' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:56.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:56 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/958667374' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:56.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:56] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:56.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:56] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:56.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:56.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:56] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:56.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:56] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:56.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:56] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:56.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:56] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:56.997 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:56 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug Change detected - internal 2 / xattr 3 refreshing 2026-03-10T05:34:57.148 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.148 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.148 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:56] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:57.148 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:56] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:57.148 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.148 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.148 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.148 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.148 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:56] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:57.148 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:56] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:57.148 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:56] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:57.148 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:56] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:56] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:56 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug gateway validation needed for vm05.local 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _gateway update on vm05.local, successful 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug An additional 1 tpg's are required 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Gateway.create_tpg) created TPG '2' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _gateway update on vm01.local, successful 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:57] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-10T05:34:57.149 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:57] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-10T05:34:57.149 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-mon[47941]: pgmap v221: 4 pgs: 4 active+clean; 450 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 28 KiB/s rd, 597 B/s wr, 35 op/s 2026-03-10T05:34:57.149 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/757685186' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:57.149 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/540427811' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:57.149 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/500184790' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:57.149 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1367093309' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:57.149 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/290550720' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:57.149 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4077812148' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:57.149 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/856176877' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:57.149 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3628604202' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:57.149 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3141362326' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:57.149 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1109578649' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:57.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:56 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:56] "GET /api/sysinfo/hostname HTTP/1.1" 200 - 2026-03-10T05:34:57.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:56 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:56] "GET /api/sysinfo/hostname HTTP/1.1" 200 - 2026-03-10T05:34:57.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-10T05:34:57.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-10T05:34:57.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-10T05:34:57.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-10T05:34:57.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-10T05:34:57.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-10T05:34:57.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:57.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:57.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-10T05:34:57.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Gateway.create_tpg) created TPG '1' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-10T05:34:57.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Gateway.create_tpg) created TPG '2' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-10T05:34:57.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Gateway.create_target) created an iscsi target with iqn of 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-10T05:34:57.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug iqn.2003-01.com.redhat.iscsi-gw:ceph-gw - Adding the IP to the enabled tpg, allowing iSCSI logins 2026-03-10T05:34:57.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-10T05:34:57.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-10T05:34:57.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:57.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:57.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:57 vm05 ceph-mon[50927]: pgmap v221: 4 pgs: 4 active+clean; 450 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 28 KiB/s rd, 597 B/s wr, 35 op/s 2026-03-10T05:34:57.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:57 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/757685186' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:57.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:57 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/540427811' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:57.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:57 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/500184790' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:57.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:57 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1367093309' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:57.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:57 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/290550720' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:57.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:57 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4077812148' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:57.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:57 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/856176877' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:57.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:57 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3628604202' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:57.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:57 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3141362326' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:57.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:57 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1109578649' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:57.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:57 vm02 ceph-mon[50473]: pgmap v221: 4 pgs: 4 active+clean; 450 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 28 KiB/s rd, 597 B/s wr, 35 op/s 2026-03-10T05:34:57.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:57 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/757685186' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:57.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:57 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/540427811' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:57.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:57 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/500184790' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:57.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:57 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1367093309' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:57.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:57 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/290550720' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:57.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:57 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4077812148' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:57.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:57 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/856176877' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:57.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:57 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3628604202' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:57.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:57 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3141362326' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:57.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:57 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1109578649' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:57.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:57] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:57] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:57] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:57.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:57] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:57.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:57] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:57] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:57] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:57] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:57.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:57.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:57.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:57.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.050 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.050 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.050 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.050 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:57 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:57] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.229 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:57] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:58.229 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:57 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:57] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:58.229 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:58] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:58.229 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:58] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:58.229 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:58] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:58.229 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:58] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:58.229 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:58] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:58.230 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:58] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:58.230 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.230 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.230 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.230 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.230 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.230 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.230 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block0' to LIO backstore user:rbd 2026-03-10T05:34:58.230 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: /usr/lib/python3.9/site-packages/rtslib_fb/root.py:180: UserWarning: Cannot set dbroot to /var/target. Target devices have already been registered. 2026-03-10T05:34:58.230 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: warn("Cannot set dbroot to {}. Target devices have already been registered." 2026-03-10T05:34:58.230 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (LUN.add_dev_to_lio) Successfully added datapool/block0 to LIO 2026-03-10T05:34:58.230 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1053864514' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:58.230 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1497917399' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:58.230 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1731517890' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:58.230 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4206158153' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:58.230 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/195991808' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:58.230 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/939911183' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:58.230 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3037032203' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:58.230 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/782135354' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:58.230 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1425753941' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:58.230 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1365447999' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:58.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:58.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:58 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1053864514' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:58.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:58 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1497917399' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:58.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:58 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1731517890' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:58.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:58 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4206158153' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:58.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:58 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/195991808' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:58.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:58 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/939911183' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:58.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:58 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3037032203' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:58.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:58 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/782135354' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:58.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:58 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1425753941' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:58.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:58 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1365447999' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:58.488 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-10T05:34:58.488 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Setup group ano2 for datapool.block0 on tpg 2 (state 1, owner False, failover type 1) 2026-03-10T05:34:58.488 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Setup group ao for datapool.block0 on tpg 1 (state 0, owner True, failover type 1) 2026-03-10T05:34:58.488 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:34:58.488 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:34:58.488 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _targetlun update on vm01.local, successful 2026-03-10T05:34:58.488 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _targetlun update on vm05.local, successful 2026-03-10T05:34:58.488 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:58] "PUT /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:34:58.488 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:58] "PUT /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:34:58.488 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:58] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:58.488 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:58] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:58.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:58 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1053864514' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:58.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:58 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1497917399' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:58.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:58 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1731517890' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:58.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:58 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4206158153' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:58.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:58 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/195991808' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:58.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:58 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/939911183' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:58.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:58 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3037032203' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:58.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:58 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/782135354' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:58.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:58 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1425753941' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:58.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:58 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1365447999' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:58.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block0' to LIO backstore user:rbd 2026-03-10T05:34:58.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: /usr/lib/python3.9/site-packages/rtslib_fb/root.py:180: UserWarning: Cannot set dbroot to /var/target. Target devices have already been registered. 2026-03-10T05:34:58.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: warn("Cannot set dbroot to {}. Target devices have already been registered." 2026-03-10T05:34:58.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (LUN.add_dev_to_lio) Successfully added datapool/block0 to LIO 2026-03-10T05:34:58.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (LUN.allocate) added block0 to LIO using wwn '36001405da17b74481464e9fa968746d3' defined by vm01.local 2026-03-10T05:34:58.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-10T05:34:58.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug Setup group ano2 for datapool.block0 on tpg 2 (state 1, owner False, failover type 1) 2026-03-10T05:34:58.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug Setup group ao for datapool.block0 on tpg 1 (state 0, owner True, failover type 1) 2026-03-10T05:34:58.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:34:58.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:34:59.089 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:58] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.089 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:58] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.089 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:58] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:59.089 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:58] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:59.089 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:58] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.089 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:58] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.089 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:58] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.089 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:58] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.089 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.089 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.089 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.089 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:58 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.318 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.319 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.319 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:58 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:58] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:59] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:59] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:59] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:59] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:59] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:59] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:59] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:59] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:59] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:59] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:59] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:59] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:59] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:59] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:59] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:59] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-mon[47941]: pgmap v222: 4 pgs: 4 active+clean; 450 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 597 B/s wr, 35 op/s 2026-03-10T05:34:59.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4243603191' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:34:59.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4243603191' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:34:59.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4243603191' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:34:59.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4243603191' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/81429086"}]: dispatch 2026-03-10T05:34:59.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-mon[47941]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/81429086"}]: dispatch 2026-03-10T05:34:59.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/2483483664' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:34:59.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/2483483664' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:34:59.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/2483483664' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:34:59.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1636357541' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:59.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2411310816' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:59.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3699618859' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:59.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4223431215' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:59.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3633929860' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:59 vm02 ceph-mon[50473]: pgmap v222: 4 pgs: 4 active+clean; 450 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 597 B/s wr, 35 op/s 2026-03-10T05:34:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:59 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4243603191' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:34:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:59 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4243603191' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:34:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:59 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4243603191' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:34:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:59 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4243603191' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/81429086"}]: dispatch 2026-03-10T05:34:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:59 vm02 ceph-mon[50473]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/81429086"}]: dispatch 2026-03-10T05:34:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:59 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/2483483664' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:34:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:59 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/2483483664' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:34:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:59 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/2483483664' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:34:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:59 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1636357541' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:59 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2411310816' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:59 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3699618859' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:59 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4223431215' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:59.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:34:59 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3633929860' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:59.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:59 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:59] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:59 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:59] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.806 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:59 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:59] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.806 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:59 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:59] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.806 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:59 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:59] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.806 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:59 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:59] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:34:59.806 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:59 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Client.define_client) iqn.1994-05.com.redhat:client added successfully 2026-03-10T05:34:59.806 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:59 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-10T05:34:59.806 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:59 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:59] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:34:59.806 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:34:59 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:34:59] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:34:59.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:59 vm05 ceph-mon[50927]: pgmap v222: 4 pgs: 4 active+clean; 450 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 597 B/s wr, 35 op/s 2026-03-10T05:34:59.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:59 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4243603191' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:34:59.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:59 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4243603191' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:34:59.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:59 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4243603191' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:34:59.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:59 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4243603191' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/81429086"}]: dispatch 2026-03-10T05:34:59.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:59 vm05 ceph-mon[50927]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/81429086"}]: dispatch 2026-03-10T05:34:59.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:59 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/2483483664' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:34:59.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:59 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/2483483664' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:34:59.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:59 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/2483483664' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:34:59.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:59 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1636357541' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:34:59.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:59 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2411310816' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:34:59.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:59 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3699618859' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:59.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:59 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4223431215' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:34:59.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:34:59 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3633929860' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:34:59.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Client.define_client) iqn.1994-05.com.redhat:client added successfully 2026-03-10T05:34:59.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-10T05:34:59.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:59] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:34:59.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:59] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:34:59.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _client update on localhost, successful 2026-03-10T05:34:59.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _client update on vm05.local, successful 2026-03-10T05:34:59.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:59] "PUT /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:34:59.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:59] "PUT /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:34:59.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:59] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:34:59.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:59] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:00.277 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2946007328' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:00.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:34:59] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:00.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:34:59 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:34:59] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:00.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:00] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:00.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:00] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:00.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:00] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:00.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:00] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:00.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:00] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:00.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:00] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:00.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.277 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:00 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:00 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:00 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:00 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:00 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2946007328' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:00 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1713223297' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:00 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1238526775' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:00 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2303199472' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:00 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4148629734' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:00 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2865818586' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:00 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2306555017' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:00 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1649822430' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:00 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2933619246' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:00.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:00 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3570551372' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:00.531 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1713223297' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:00.531 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1238526775' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:00.531 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2303199472' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:00.531 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4148629734' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:00.531 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2865818586' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:00.531 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2306555017' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:00.531 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1649822430' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:00.531 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2933619246' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:00.531 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3570551372' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:00.531 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:00] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:00.531 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:00] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:00.531 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:00] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:00.531 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:00] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:00.531 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:00] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:00.531 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:00] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:00.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:00 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2946007328' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:00.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:00 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1713223297' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:00.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:00 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1238526775' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:00.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:00 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2303199472' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:00.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:00 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4148629734' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:00.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:00 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2865818586' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:00.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:00 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2306555017' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:00.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:00 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1649822430' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:00.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:00 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2933619246' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:00.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:00 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3570551372' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:00.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:00 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:00 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:00 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:00 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:00] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:00.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:00] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:00.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:00.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:00] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.279 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:00] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:01.279 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:00 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:00] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:01.279 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:01] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:01.279 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:01] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:01.279 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:01] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:01] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:01] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:01] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:01] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:01] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Client.add_lun) added image 'datapool/block0' to iqn.1994-05.com.redhat:client 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:01] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:01] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _clientlun update on localhost, successful 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _clientlun update on vm05.local, successful 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:01] "PUT /api/clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:01] "PUT /api/clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:01] "GET /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:01] "GET /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:01 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:01 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:01 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:01 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:01 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:01 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:01 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Client.add_lun) added image 'datapool/block0' to iqn.1994-05.com.redhat:client 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:01 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:01 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:01 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:35:01.280 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-mon[47941]: pgmap v223: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 78 KiB/s rd, 2.0 KiB/s wr, 87 op/s 2026-03-10T05:35:01.280 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2056922423' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:01.280 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/376767534' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:01.280 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3466151375' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:01.280 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/131252602' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:01.280 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3664279244' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:01.280 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3365907092' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:01.280 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1088986835' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:01.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:01 vm05 ceph-mon[50927]: pgmap v223: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 78 KiB/s rd, 2.0 KiB/s wr, 87 op/s 2026-03-10T05:35:01.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:01 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2056922423' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:01.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:01 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/376767534' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:01.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:01 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3466151375' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:01.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:01 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/131252602' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:01.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:01 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3664279244' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:01.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:01 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3365907092' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:01.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:01 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1088986835' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:01.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:01 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2350728365' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:01.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:01 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3070085363' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:01.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:01 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1019645120' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:01.569 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2350728365' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:01.569 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3070085363' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:01.569 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1019645120' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:01.569 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:01] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:01.569 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:01] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:01.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:01 vm02 ceph-mon[50473]: pgmap v223: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 78 KiB/s rd, 2.0 KiB/s wr, 87 op/s 2026-03-10T05:35:01.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:01 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2056922423' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:01.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:01 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/376767534' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:01.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:01 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3466151375' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:01.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:01 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/131252602' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:01.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:01 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3664279244' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:01.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:01 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3365907092' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:01.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:01 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1088986835' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:01.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:01 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2350728365' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:01.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:01 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3070085363' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:01.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:01 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1019645120' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:01.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:01] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:01.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:01] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:01.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:01] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:01.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:01] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:01.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:01] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:01.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:01] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:01.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:01.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:02.055 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:01 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:02.055 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:01 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:02.055 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:01 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:02.055 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:01 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:02.124 INFO:tasks.cram.client.0.vm01.stdout:/home/ubuntu/cephtest/archive/cram.client.0/gwcli_create.t: passed 2026-03-10T05:35:02.124 INFO:tasks.cram.client.0.vm01.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-10T05:35:02.128 INFO:tasks.cram:Running tests for client.1... 2026-03-10T05:35:02.128 DEBUG:teuthology.orchestra.run.vm02:> CEPH_REF=master CEPH_ID="1" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.1/*.t 2026-03-10T05:35:02.287 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:01] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:02.287 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:01] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:02.287 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:02] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:02.287 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:02] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:02.287 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:02] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:02.287 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:02] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:02.287 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:35:02] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:02.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:35:02] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:02.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:02.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:02.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:02.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:02.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:02 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:02.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:02 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:02.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:02 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:02.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:02 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:35:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:02 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:02 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3810429999' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:02 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1372727525' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:02 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1813378595' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:02 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1865629952' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:02 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/766411807' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:02 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/1960794090' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:02 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/3110057080' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:02 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/829634777' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:02 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2418009716' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:02.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:02 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2499662422' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:02.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:02.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3810429999' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:02.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1372727525' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:02.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1813378595' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:02.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1865629952' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:02.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/766411807' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:02.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/1960794090' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:02.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/3110057080' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:02.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/829634777' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:02.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2418009716' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:02.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:02 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2499662422' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:02.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:02 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:02.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:02 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3810429999' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:02.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:02 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1372727525' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:02.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:02 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1813378595' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:02.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:02 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1865629952' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:02.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:02 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/766411807' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:02.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:02 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/1960794090' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:02.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:02 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/3110057080' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:02.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:02 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/829634777' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:02.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:02 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2418009716' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:02.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:02 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2499662422' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:03.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:03 vm05 ceph-mon[50927]: pgmap v224: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 79 KiB/s rd, 2.0 KiB/s wr, 90 op/s 2026-03-10T05:35:03.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:03 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:03.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:03 vm01 ceph-mon[47941]: pgmap v224: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 79 KiB/s rd, 2.0 KiB/s wr, 90 op/s 2026-03-10T05:35:03.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:03 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:03.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:03 vm02 ceph-mon[50473]: pgmap v224: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 79 KiB/s rd, 2.0 KiB/s wr, 90 op/s 2026-03-10T05:35:03.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:03 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:05.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:05 vm05 ceph-mon[50927]: pgmap v225: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 79 KiB/s rd, 1.9 KiB/s wr, 89 op/s 2026-03-10T05:35:05.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:05 vm01 ceph-mon[47941]: pgmap v225: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 79 KiB/s rd, 1.9 KiB/s wr, 89 op/s 2026-03-10T05:35:05.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:05 vm02 ceph-mon[50473]: pgmap v225: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 79 KiB/s rd, 1.9 KiB/s wr, 89 op/s 2026-03-10T05:35:07.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:07 vm05 ceph-mon[50927]: pgmap v226: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 99 KiB/s rd, 2.2 KiB/s wr, 107 op/s 2026-03-10T05:35:07.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:07 vm01 ceph-mon[47941]: pgmap v226: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 99 KiB/s rd, 2.2 KiB/s wr, 107 op/s 2026-03-10T05:35:07.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:07 vm02 ceph-mon[50473]: pgmap v226: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 99 KiB/s rd, 2.2 KiB/s wr, 107 op/s 2026-03-10T05:35:09.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:09 vm01 ceph-mon[47941]: pgmap v227: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 73 KiB/s rd, 1.7 KiB/s wr, 73 op/s 2026-03-10T05:35:09.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:09 vm02 ceph-mon[50473]: pgmap v227: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 73 KiB/s rd, 1.7 KiB/s wr, 73 op/s 2026-03-10T05:35:09.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:09 vm05 ceph-mon[50927]: pgmap v227: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 73 KiB/s rd, 1.7 KiB/s wr, 73 op/s 2026-03-10T05:35:11.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:11 vm01 ceph-mon[47941]: pgmap v228: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 73 KiB/s rd, 1.7 KiB/s wr, 74 op/s 2026-03-10T05:35:11.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:11 vm02 ceph-mon[50473]: pgmap v228: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 73 KiB/s rd, 1.7 KiB/s wr, 74 op/s 2026-03-10T05:35:11.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:11 vm05 ceph-mon[50927]: pgmap v228: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 73 KiB/s rd, 1.7 KiB/s wr, 74 op/s 2026-03-10T05:35:12.487 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:12 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:12.534 INFO:tasks.cram.client.1.vm02.stdout:/home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t: passed 2026-03-10T05:35:12.535 INFO:tasks.cram.client.1.vm02.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-10T05:35:12.536 INFO:tasks.cram:Running tests for client.2... 2026-03-10T05:35:12.536 DEBUG:teuthology.orchestra.run.vm05:> CEPH_REF=master CEPH_ID="2" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.2/*.t 2026-03-10T05:35:12.582 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:12 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:12.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:12 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:13.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:12 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:12] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:13.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:12 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:12] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:13.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:13] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:13.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:13] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:13.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:13] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:13.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:13] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:13.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:13] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:13.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:13] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:13.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:13] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:13.306 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:13] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:13.330 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.330 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.330 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.330 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-mon[47941]: pgmap v229: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 341 B/s wr, 22 op/s 2026-03-10T05:35:13.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:13.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/955632129' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:13.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3817231969' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:13.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/2793652208' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:13.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1212343878' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:13.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3426945918' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:13.587 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:13] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:13] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:13] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:13] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:13] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:13] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:13] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:13] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:13] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:13] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:13] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:13.594 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:13 vm05 ceph-mon[50927]: pgmap v229: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 341 B/s wr, 22 op/s 2026-03-10T05:35:13.594 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:13 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:13.594 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:13 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/955632129' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:13.594 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:13 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3817231969' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:13.594 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:13 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/2793652208' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:13.594 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:13 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1212343878' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:13.594 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:13 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3426945918' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:13.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:13 vm02 ceph-mon[50473]: pgmap v229: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 341 B/s wr, 22 op/s 2026-03-10T05:35:13.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:13 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:13.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:13 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/955632129' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:13.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:13 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3817231969' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:13.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:13 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/2793652208' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:13.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:13 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1212343878' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:13.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:13 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3426945918' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:14.006 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug _client update on vm01.local, successful 2026-03-10T05:35:14.006 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Client.delete) deleted NodeACL for iqn.1994-05.com.redhat:client 2026-03-10T05:35:14.006 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:13] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:35:14.006 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:13] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:35:14.006 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug _client update on localhost, successful 2026-03-10T05:35:14.006 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:13] "DELETE /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:35:14.006 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:13 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:13] "DELETE /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:35:14.087 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Client.delete) deleted NodeACL for iqn.1994-05.com.redhat:client 2026-03-10T05:35:14.087 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:35:14.087 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:13 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:13] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:35:14.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:14] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:14.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:14] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:14.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:14] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:14.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:14] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:14.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:14] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:14.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:14] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:14.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:14] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:14.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:14] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:14.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.305 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.543 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3797799324' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:14.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1288028041' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:14.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3778917510' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:14.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/668506808' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:14.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1640325842' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:14.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3717585637' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:14.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1482084626' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:14.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3208220965' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:14.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/2826419212' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:14.544 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/749231260' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:14.544 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.544 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.544 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.544 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.544 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Change detected - internal 8 / xattr 9 refreshing 2026-03-10T05:35:14.597 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:14 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3797799324' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:14.597 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:14 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1288028041' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:14.597 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:14 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3778917510' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:14.597 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:14 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/668506808' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:14.597 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:14 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1640325842' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:14.597 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:14 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3717585637' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:14.598 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:14 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1482084626' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:14.598 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:14 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3208220965' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:14.598 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:14 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/2826419212' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:14.598 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:14 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/749231260' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:14.598 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:14] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:14.598 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:14] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:14.598 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:14] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:14.598 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:14] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:14.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:14 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3797799324' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:14.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:14 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1288028041' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:14.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:14 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3778917510' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:14.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:14 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/668506808' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:14.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:14 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1640325842' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:14.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:14 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3717585637' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:14.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:14 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1482084626' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:14.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:14 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3208220965' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:14.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:14 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/2826419212' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:14.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:14 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/749231260' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:14.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug LUN unmap request received, config commit to be performed by vm05.local 2026-03-10T05:35:14.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:35:14.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:14 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:35:14.897 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:14] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:14.897 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:14] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:14.897 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:14] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:14.897 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:14] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:14.897 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.897 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.898 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.898 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.898 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.898 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:14.898 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug _targetlun update on vm01.local, successful 2026-03-10T05:35:14.898 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug LUN unmap request received, config commit to be performed by vm05.local 2026-03-10T05:35:14.898 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:35:14.898 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:14] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:35:14.898 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug _targetlun update on vm05.local, successful 2026-03-10T05:35:14.898 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:14] "DELETE /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:35:14.898 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:14 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:14] "DELETE /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:35:15.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:15] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:15.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:15] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:15.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:15] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:15.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:15] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:15.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:15] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:15.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:15] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:15.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:15] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:15.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:15] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:15.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:15.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:15.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:15.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:15.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:15 vm05 ceph-mon[50927]: pgmap v230: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 21 KiB/s rd, 341 B/s wr, 17 op/s 2026-03-10T05:35:15.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:35:15.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:15 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/230036709' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:15.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:15 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/47644424' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:15.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:15 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/999956699' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:15.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:15 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3323871856' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:15.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:15 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3788615779' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:15.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:35:15.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:35:15.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:35:15.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:15 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1535873774' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:15.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:15 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3621861256' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:15.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:15 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/2711079038' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:15.587 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:15.587 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:15.587 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:15.587 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:15.587 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Change detected - internal 9 / xattr 10 refreshing 2026-03-10T05:35:15.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-mon[47941]: pgmap v230: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 21 KiB/s rd, 341 B/s wr, 17 op/s 2026-03-10T05:35:15.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:35:15.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/230036709' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:15.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/47644424' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:15.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/999956699' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:15.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3323871856' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:15.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3788615779' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:15.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:35:15.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:35:15.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:35:15.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1535873774' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:15.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3621861256' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:15.588 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/2711079038' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:15 vm02 ceph-mon[50473]: pgmap v230: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 21 KiB/s rd, 341 B/s wr, 17 op/s 2026-03-10T05:35:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:35:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:15 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/230036709' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:15 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/47644424' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:15 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/999956699' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:15 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3323871856' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:15 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3788615779' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:35:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:35:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:35:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:15 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1535873774' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:15 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3621861256' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:15.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:15 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/2711079038' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:15.997 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:15] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:15] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:15] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:15] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:15] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:15] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:15] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:15] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug _target update on vm01.local, successful 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug Removing target configuration 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug _target update on vm05.local, successful 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:15] "DELETE /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:35:15.998 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:15 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:15] "DELETE /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:35:16.087 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:16.088 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:16.088 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:16.088 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:16.088 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:16.088 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:16.088 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-10T05:35:16.088 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Removing target configuration 2026-03-10T05:35:16.088 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:35:16.088 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:15 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:15] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:35:16.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:16 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/4266952942' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:16.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:16 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3701555543' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:16.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:16 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3670307613' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:16.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:16 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3264315274' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:16.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:16 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3223128107' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:16.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:16 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1970030231' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:16.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:16 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1121699641' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:16.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:16 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/2359348490' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:16.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:16 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3751752080' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:16.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:16 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:16] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:16.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:16 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:16] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:16.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:16 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:16] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:16.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:16 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:16] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:16.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:16 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/4266952942' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:16.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:16 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3701555543' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:16.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:16 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3670307613' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:16.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:16 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3264315274' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:16.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:16 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3223128107' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:16.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:16 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1970030231' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:16.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:16 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1121699641' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:16.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:16 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/2359348490' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:16.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:16 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3751752080' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:16.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:16 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Change detected - internal 11 / xattr 12 refreshing 2026-03-10T05:35:16.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:16 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/4266952942' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:16.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:16 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3701555543' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:16.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:16 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3670307613' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:16.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:16 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3264315274' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:16.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:16 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3223128107' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:16.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:16 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1970030231' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:16.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:16 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1121699641' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:16.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:16 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/2359348490' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:16.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:16 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3751752080' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:17.089 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:16 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:16] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:17.089 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:16 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:16] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:17.089 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:16 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:16] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:17.089 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:16 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:16] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:17.089 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:16 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:17.089 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:16 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:16] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:35:17.089 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:16 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug LUN deletion request received, rbd removal to be performed by vm05.local 2026-03-10T05:35:17.089 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:17 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:17] "DELETE /api/_disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:17.089 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:17 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:17] "DELETE /api/_disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:17.089 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:17 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug _disk update on vm05.local, successful 2026-03-10T05:35:17.089 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:17 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:17] "DELETE /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:17.089 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:17 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:35:17] "DELETE /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-10T05:35:17.519 INFO:tasks.cram.client.2.vm05.stdout:/home/ubuntu/cephtest/archive/cram.client.2/gwcli_delete.t: passed 2026-03-10T05:35:17.519 INFO:tasks.cram.client.2.vm05.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-10T05:35:17.522 DEBUG:teuthology.orchestra.run.vm01:> test -f /home/ubuntu/cephtest/archive/cram.client.0/gwcli_create.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.0/gwcli_create.t 2026-03-10T05:35:17.538 DEBUG:teuthology.orchestra.run.vm01:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.0 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.0 2026-03-10T05:35:17.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:17 vm02 ceph-mon[50473]: pgmap v231: 4 pgs: 4 active+clean; 11 MiB data, 217 MiB used, 160 GiB / 160 GiB avail; 129 KiB/s rd, 38 KiB/s wr, 54 op/s 2026-03-10T05:35:17.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:17 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/715046522' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:17.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:17 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/2508722015' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:17.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:17 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3158711613' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:17.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:17 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/29397984' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:17.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:17 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/759611274' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:17.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:17 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/2709533604' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:17.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:17 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/59358055' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:17.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:17 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/2986445450' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:17.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:17 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/2905166016' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:17.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:17 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:35:17] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:17.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:17 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:35:17] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:35:17.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:17 vm05 ceph-mon[50927]: pgmap v231: 4 pgs: 4 active+clean; 11 MiB data, 217 MiB used, 160 GiB / 160 GiB avail; 129 KiB/s rd, 38 KiB/s wr, 54 op/s 2026-03-10T05:35:17.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:17 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/715046522' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:17.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:17 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/2508722015' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:17.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:17 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3158711613' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:17.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:17 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/29397984' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:17.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:17 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/759611274' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:17.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:17 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/2709533604' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:17.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:17 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/59358055' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:17.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:17 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/2986445450' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:17.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:17 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/2905166016' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:17.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:17 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Change detected - internal 12 / xattr 13 refreshing 2026-03-10T05:35:17.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:17 vm01 ceph-mon[47941]: pgmap v231: 4 pgs: 4 active+clean; 11 MiB data, 217 MiB used, 160 GiB / 160 GiB avail; 129 KiB/s rd, 38 KiB/s wr, 54 op/s 2026-03-10T05:35:17.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:17 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/715046522' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:17.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:17 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/2508722015' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:17.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:17 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3158711613' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:17.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:17 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/29397984' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:17.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:17 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/759611274' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:17.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:17 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/2709533604' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:17.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:17 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/59358055' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:17.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:17 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/2986445450' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:17.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:17 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/2905166016' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:17.950 DEBUG:teuthology.orchestra.run.vm02:> test -f /home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t 2026-03-10T05:35:17.964 DEBUG:teuthology.orchestra.run.vm02:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.1 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.1 2026-03-10T05:35:18.365 DEBUG:teuthology.orchestra.run.vm05:> test -f /home/ubuntu/cephtest/archive/cram.client.2/gwcli_delete.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.2/gwcli_delete.t 2026-03-10T05:35:18.380 DEBUG:teuthology.orchestra.run.vm05:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.2 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.2 2026-03-10T05:35:18.737 INFO:teuthology.run_tasks:Running task cram... 2026-03-10T05:35:18.739 INFO:tasks.cram:Pulling tests from https://github.com/kshtsk/ceph.git ref 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b 2026-03-10T05:35:18.740 DEBUG:teuthology.orchestra.run.vm01:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.0 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-10T05:35:18.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:18 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/4097736283' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:18.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:18 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/875681050' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:18.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:18 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1146864176' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:18.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:18 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1146920219' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:18.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:18 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/3545683188' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:18.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:18 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/4097736283' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:18.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:18 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/875681050' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:18.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:18 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1146864176' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:18.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:18 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1146920219' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:18.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:18 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/3545683188' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:18.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:18 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/4097736283' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "version"}]: dispatch 2026-03-10T05:35:18.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:18 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/875681050' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T05:35:18.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:18 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1146864176' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:18.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:18 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1146920219' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-10T05:35:18.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:18 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/3545683188' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "df", "format": "json"}]: dispatch 2026-03-10T05:35:19.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:19 vm05 ceph-mon[50927]: pgmap v232: 4 pgs: 4 active+clean; 11 MiB data, 217 MiB used, 160 GiB / 160 GiB avail; 109 KiB/s rd, 38 KiB/s wr, 37 op/s 2026-03-10T05:35:19.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:19 vm01 ceph-mon[47941]: pgmap v232: 4 pgs: 4 active+clean; 11 MiB data, 217 MiB used, 160 GiB / 160 GiB avail; 109 KiB/s rd, 38 KiB/s wr, 37 op/s 2026-03-10T05:35:19.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:19 vm02 ceph-mon[50473]: pgmap v232: 4 pgs: 4 active+clean; 11 MiB data, 217 MiB used, 160 GiB / 160 GiB avail; 109 KiB/s rd, 38 KiB/s wr, 37 op/s 2026-03-10T05:35:20.118 INFO:teuthology.orchestra.run.vm01.stdout:Collecting cram==0.6 2026-03-10T05:35:20.119 INFO:teuthology.orchestra.run.vm01.stdout: Using cached cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-10T05:35:20.129 INFO:teuthology.orchestra.run.vm01.stdout:Installing collected packages: cram 2026-03-10T05:35:20.135 INFO:teuthology.orchestra.run.vm01.stdout:Successfully installed cram-0.6 2026-03-10T05:35:20.136 INFO:teuthology.orchestra.run.vm01.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-03-10T05:35:20.136 INFO:teuthology.orchestra.run.vm01.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/virtualenv/bin/python3 -m pip install --upgrade pip' command. 2026-03-10T05:35:20.170 DEBUG:teuthology.orchestra.run.vm01:> rm -rf /home/ubuntu/cephtest/clone.client.0 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.0 && cd /home/ubuntu/cephtest/clone.client.0 && git checkout 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b 2026-03-10T05:35:20.187 INFO:teuthology.orchestra.run.vm01.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.0'... 2026-03-10T05:35:21.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:21 vm05 ceph-mon[50927]: pgmap v233: 4 pgs: 4 active+clean; 4.6 MiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 65 op/s 2026-03-10T05:35:21.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:35:21.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:21 vm01 ceph-mon[47941]: pgmap v233: 4 pgs: 4 active+clean; 4.6 MiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 65 op/s 2026-03-10T05:35:21.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:21 vm02 ceph-mon[50473]: pgmap v233: 4 pgs: 4 active+clean; 4.6 MiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 65 op/s 2026-03-10T05:35:22.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:22 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:35:22.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:22 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:22.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:22 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:22.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:22 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:23.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:23 vm05 ceph-mon[50927]: pgmap v234: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 66 op/s 2026-03-10T05:35:23.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:23 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:23.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:23 vm01 ceph-mon[47941]: pgmap v234: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 66 op/s 2026-03-10T05:35:23.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:23 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:23.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:23 vm02 ceph-mon[50473]: pgmap v234: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 66 op/s 2026-03-10T05:35:23.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:23 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:25.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:25 vm05 ceph-mon[50927]: pgmap v235: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 66 op/s 2026-03-10T05:35:25.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:25 vm01 ceph-mon[47941]: pgmap v235: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 66 op/s 2026-03-10T05:35:25.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:25 vm02 ceph-mon[50473]: pgmap v235: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 133 KiB/s rd, 39 KiB/s wr, 66 op/s 2026-03-10T05:35:27.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:27 vm05 ceph-mon[50927]: pgmap v236: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 134 KiB/s rd, 39 KiB/s wr, 67 op/s 2026-03-10T05:35:27.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:27 vm01 ceph-mon[47941]: pgmap v236: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 134 KiB/s rd, 39 KiB/s wr, 67 op/s 2026-03-10T05:35:27.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:27 vm02 ceph-mon[50473]: pgmap v236: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 134 KiB/s rd, 39 KiB/s wr, 67 op/s 2026-03-10T05:35:29.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:29 vm05 ceph-mon[50927]: pgmap v237: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 25 KiB/s rd, 1.2 KiB/s wr, 30 op/s 2026-03-10T05:35:29.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:29 vm01 ceph-mon[47941]: pgmap v237: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 25 KiB/s rd, 1.2 KiB/s wr, 30 op/s 2026-03-10T05:35:29.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:29 vm02 ceph-mon[50473]: pgmap v237: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 25 KiB/s rd, 1.2 KiB/s wr, 30 op/s 2026-03-10T05:35:31.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:31 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:35:32.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:31 vm02 ceph-mon[50473]: pgmap v238: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 26 KiB/s rd, 1.2 KiB/s wr, 31 op/s 2026-03-10T05:35:32.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:31 vm05 ceph-mon[50927]: pgmap v238: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 26 KiB/s rd, 1.2 KiB/s wr, 31 op/s 2026-03-10T05:35:32.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:31 vm01 ceph-mon[47941]: pgmap v238: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 26 KiB/s rd, 1.2 KiB/s wr, 31 op/s 2026-03-10T05:35:32.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:32 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:35:33.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:32 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:33.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:32 vm02 ceph-mon[50473]: pgmap v239: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 255 B/s wr, 3 op/s 2026-03-10T05:35:33.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:32 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:33.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:32 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:33.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:32 vm05 ceph-mon[50927]: pgmap v239: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 255 B/s wr, 3 op/s 2026-03-10T05:35:33.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:32 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:33.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:32 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:33.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:32 vm01 ceph-mon[47941]: pgmap v239: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 255 B/s wr, 3 op/s 2026-03-10T05:35:33.338 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:32 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:35.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:34 vm05 ceph-mon[50927]: pgmap v240: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:35.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:34 vm01 ceph-mon[47941]: pgmap v240: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:35.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:34 vm02 ceph-mon[50473]: pgmap v240: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:37.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:36 vm05 ceph-mon[50927]: pgmap v241: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:37.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:36 vm01 ceph-mon[47941]: pgmap v241: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:37.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:36 vm02 ceph-mon[50473]: pgmap v241: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:39.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:38 vm05 ceph-mon[50927]: pgmap v242: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:39.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:38 vm01 ceph-mon[47941]: pgmap v242: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:39.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:38 vm02 ceph-mon[50473]: pgmap v242: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:41.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:40 vm05 ceph-mon[50927]: pgmap v243: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:41.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:40 vm01 ceph-mon[47941]: pgmap v243: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:41.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:40 vm02 ceph-mon[50473]: pgmap v243: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:41.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:41 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:35:42.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:41 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:42.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:41 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:42.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:41 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:42.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:42 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:35:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:42 vm05 ceph-mon[50927]: pgmap v244: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:43.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:42 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:43.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:42 vm01 ceph-mon[47941]: pgmap v244: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:43.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:42 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:42 vm02 ceph-mon[50473]: pgmap v244: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:43.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:42 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:45.262 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:44 vm05 ceph-mon[50927]: pgmap v245: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:45.344 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:44 vm01 ceph-mon[47941]: pgmap v245: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:45.350 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:44 vm02 ceph-mon[50473]: pgmap v245: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:47.088 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:47 vm01 ceph-mon[47941]: pgmap v246: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:47.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:47 vm05 ceph-mon[50927]: pgmap v246: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:47.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:47 vm02 ceph-mon[50473]: pgmap v246: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:49.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:49 vm05 ceph-mon[50927]: pgmap v247: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:49.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:49 vm01 ceph-mon[47941]: pgmap v247: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:49.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:49 vm02 ceph-mon[50473]: pgmap v247: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:51.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:51 vm05 ceph-mon[50927]: pgmap v248: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:51.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:51 vm01 ceph-mon[47941]: pgmap v248: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:51.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:51 vm02 ceph-mon[50473]: pgmap v248: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:51.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:35:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:35:52.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:52 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:52.484 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:52 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:52.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:52 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:52.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:35:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:35:53.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:53 vm02 ceph-mon[50473]: pgmap v249: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:53.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:53 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:53 vm05 ceph-mon[50927]: pgmap v249: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:53 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:53.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:53 vm01 ceph-mon[47941]: pgmap v249: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:53.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:53 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:35:55.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:55 vm02 ceph-mon[50473]: pgmap v250: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:55.535 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:55 vm05 ceph-mon[50927]: pgmap v250: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:55.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:55 vm01 ceph-mon[47941]: pgmap v250: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:57.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:57 vm05 ceph-mon[50927]: pgmap v251: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:57.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:57 vm01 ceph-mon[47941]: pgmap v251: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:57.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:57 vm02 ceph-mon[50473]: pgmap v251: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:35:59.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:35:59 vm05 ceph-mon[50927]: pgmap v252: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:59.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:35:59 vm01 ceph-mon[47941]: pgmap v252: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:35:59.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:35:59 vm02 ceph-mon[50473]: pgmap v252: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:01.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:01 vm05 ceph-mon[50927]: pgmap v253: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:01.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:01 vm01 ceph-mon[47941]: pgmap v253: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:01.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:36:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:36:01.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:01 vm02 ceph-mon[50473]: pgmap v253: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:02.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:36:02 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:36:02.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:02 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:02.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:02 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:02.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:02 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:03 vm05 ceph-mon[50927]: pgmap v254: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:03 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:03.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:03 vm01 ceph-mon[47941]: pgmap v254: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:03.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:03 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:03.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:03 vm02 ceph-mon[50473]: pgmap v254: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:03.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:03 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr:Note: switching to '75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b'. 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr:state without impacting any branches by switching back to a branch. 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr: git switch -c 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr:Or undo this operation with: 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr: git switch - 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:36:05.629 INFO:teuthology.orchestra.run.vm01.stderr:HEAD is now at 75a68fd8ca3 qa/suites/orch/cephadm/osds: drop nvme_loop task 2026-03-10T05:36:05.635 DEBUG:teuthology.orchestra.run.vm01:> cp -- /home/ubuntu/cephtest/clone.client.0/src/test/cli-integration/rbd/rest_api_create.t /home/ubuntu/cephtest/archive/cram.client.0 2026-03-10T05:36:05.690 DEBUG:teuthology.orchestra.run.vm02:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.1 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-10T05:36:05.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:05 vm05 ceph-mon[50927]: pgmap v255: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:05.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:05 vm01 ceph-mon[47941]: pgmap v255: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:05.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:05 vm02 ceph-mon[50473]: pgmap v255: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:07.095 INFO:teuthology.orchestra.run.vm02.stdout:Collecting cram==0.6 2026-03-10T05:36:07.095 INFO:teuthology.orchestra.run.vm02.stdout: Using cached cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-10T05:36:07.106 INFO:teuthology.orchestra.run.vm02.stdout:Installing collected packages: cram 2026-03-10T05:36:07.112 INFO:teuthology.orchestra.run.vm02.stdout:Successfully installed cram-0.6 2026-03-10T05:36:07.113 INFO:teuthology.orchestra.run.vm02.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-03-10T05:36:07.113 INFO:teuthology.orchestra.run.vm02.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/virtualenv/bin/python3 -m pip install --upgrade pip' command. 2026-03-10T05:36:07.146 DEBUG:teuthology.orchestra.run.vm02:> rm -rf /home/ubuntu/cephtest/clone.client.1 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.1 && cd /home/ubuntu/cephtest/clone.client.1 && git checkout 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b 2026-03-10T05:36:07.161 INFO:teuthology.orchestra.run.vm02.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.1'... 2026-03-10T05:36:07.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:07 vm05 ceph-mon[50927]: pgmap v256: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:07.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:07 vm01 ceph-mon[47941]: pgmap v256: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:07.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:07 vm02 ceph-mon[50473]: pgmap v256: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:09.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:09 vm05 ceph-mon[50927]: pgmap v257: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:09.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:09 vm01 ceph-mon[47941]: pgmap v257: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:09.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:09 vm02 ceph-mon[50473]: pgmap v257: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:11.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:11 vm05 ceph-mon[50927]: pgmap v258: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:11.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:36:11 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:36:11.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:11 vm01 ceph-mon[47941]: pgmap v258: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:11.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:11 vm02 ceph-mon[50473]: pgmap v258: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:12.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:36:12 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:36:12.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:12 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:12.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:12 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:12.902 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:12 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:13.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:13 vm05 ceph-mon[50927]: pgmap v259: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:13.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:13 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:13.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:13 vm01 ceph-mon[47941]: pgmap v259: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:13.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:13 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:13.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:13 vm02 ceph-mon[50473]: pgmap v259: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:13.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:13 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:15.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:15 vm01 ceph-mon[47941]: pgmap v260: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:15.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:36:15.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:36:15.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:36:15.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:36:15.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:15 vm02 ceph-mon[50473]: pgmap v260: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:15.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:36:15.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:36:15.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:36:15.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:36:16.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:15 vm05 ceph-mon[50927]: pgmap v260: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:16.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:36:16.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:36:16.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:36:16.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:36:17.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:17 vm01 ceph-mon[47941]: pgmap v261: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:17.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:17 vm02 ceph-mon[50473]: pgmap v261: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:18.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:17 vm05 ceph-mon[50927]: pgmap v261: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:20.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:19 vm05 ceph-mon[50927]: pgmap v262: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:20.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:19 vm01 ceph-mon[47941]: pgmap v262: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:20.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:19 vm02 ceph-mon[50473]: pgmap v262: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:21.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:20 vm05 ceph-mon[50927]: pgmap v263: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:21.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:20 vm01 ceph-mon[47941]: pgmap v263: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:21.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:20 vm02 ceph-mon[50473]: pgmap v263: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:21.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:36:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:36:22.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:21 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:22.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:21 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:22.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:21 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:22.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:36:22 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:36:23.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:22 vm05 ceph-mon[50927]: pgmap v264: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:23.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:22 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:23.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:22 vm01 ceph-mon[47941]: pgmap v264: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:23.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:22 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:23.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:22 vm02 ceph-mon[50473]: pgmap v264: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:23.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:22 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:25.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:24 vm05 ceph-mon[50927]: pgmap v265: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:25.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:24 vm01 ceph-mon[47941]: pgmap v265: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:25.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:24 vm02 ceph-mon[50473]: pgmap v265: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:27.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:26 vm05 ceph-mon[50927]: pgmap v266: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:27.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:26 vm01 ceph-mon[47941]: pgmap v266: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:27.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:26 vm02 ceph-mon[50473]: pgmap v266: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:29.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:28 vm05 ceph-mon[50927]: pgmap v267: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:29.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:28 vm01 ceph-mon[47941]: pgmap v267: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:29.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:28 vm02 ceph-mon[50473]: pgmap v267: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:31.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:30 vm05 ceph-mon[50927]: pgmap v268: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:31.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:30 vm01 ceph-mon[47941]: pgmap v268: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:31.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:30 vm02 ceph-mon[50473]: pgmap v268: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:31.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:36:31 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:36:32.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:32 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:32.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:32 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:32.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:32 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:32.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:36:32 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:36:33.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:33 vm05 ceph-mon[50927]: pgmap v269: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:33.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:33 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:33.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:33 vm01 ceph-mon[47941]: pgmap v269: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:33.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:33 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:33.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:33 vm02 ceph-mon[50473]: pgmap v269: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:33.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:33 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:35.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:35 vm05 ceph-mon[50927]: pgmap v270: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:35.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:35 vm01 ceph-mon[47941]: pgmap v270: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:35.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:35 vm02 ceph-mon[50473]: pgmap v270: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:37.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:37 vm05 ceph-mon[50927]: pgmap v271: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:37.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:37 vm01 ceph-mon[47941]: pgmap v271: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:37.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:37 vm02 ceph-mon[50473]: pgmap v271: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:39.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:39 vm05 ceph-mon[50927]: pgmap v272: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:39.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:39 vm01 ceph-mon[47941]: pgmap v272: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:39.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:39 vm02 ceph-mon[50473]: pgmap v272: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:41.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:41 vm02 ceph-mon[50473]: pgmap v273: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:41.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:41 vm05 ceph-mon[50927]: pgmap v273: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:41.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:41 vm01 ceph-mon[47941]: pgmap v273: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:41.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:36:41 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:36:42.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:42 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:42.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:42 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:42.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:36:42 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:36:42.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:42 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:43.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:43 vm02 ceph-mon[50473]: pgmap v274: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:43.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:43 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:43.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:43 vm05 ceph-mon[50927]: pgmap v274: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:43.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:43 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:43.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:43 vm01 ceph-mon[47941]: pgmap v274: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:43.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:43 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:45.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:45 vm05 ceph-mon[50927]: pgmap v275: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:45.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:45 vm01 ceph-mon[47941]: pgmap v275: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:45.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:45 vm02 ceph-mon[50473]: pgmap v275: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:47.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:47 vm05 ceph-mon[50927]: pgmap v276: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:47.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:47 vm01 ceph-mon[47941]: pgmap v276: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:47.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:47 vm02 ceph-mon[50473]: pgmap v276: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr:Note: switching to '75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b'. 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr:state without impacting any branches by switching back to a branch. 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr: git switch -c 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr:Or undo this operation with: 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr: git switch - 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:36:48.615 INFO:teuthology.orchestra.run.vm02.stderr:HEAD is now at 75a68fd8ca3 qa/suites/orch/cephadm/osds: drop nvme_loop task 2026-03-10T05:36:48.620 DEBUG:teuthology.orchestra.run.vm02:> cp -- /home/ubuntu/cephtest/clone.client.1/src/test/cli-integration/rbd/iscsi_client.t /home/ubuntu/cephtest/archive/cram.client.1 2026-03-10T05:36:48.676 DEBUG:teuthology.orchestra.run.vm05:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.2 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-10T05:36:49.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:49 vm05 ceph-mon[50927]: pgmap v277: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:49.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:49 vm01 ceph-mon[47941]: pgmap v277: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:49.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:49 vm02 ceph-mon[50473]: pgmap v277: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:50.071 INFO:teuthology.orchestra.run.vm05.stdout:Collecting cram==0.6 2026-03-10T05:36:50.072 INFO:teuthology.orchestra.run.vm05.stdout: Using cached cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-10T05:36:50.084 INFO:teuthology.orchestra.run.vm05.stdout:Installing collected packages: cram 2026-03-10T05:36:50.091 INFO:teuthology.orchestra.run.vm05.stdout:Successfully installed cram-0.6 2026-03-10T05:36:50.092 INFO:teuthology.orchestra.run.vm05.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-03-10T05:36:50.092 INFO:teuthology.orchestra.run.vm05.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/virtualenv/bin/python3 -m pip install --upgrade pip' command. 2026-03-10T05:36:50.126 DEBUG:teuthology.orchestra.run.vm05:> rm -rf /home/ubuntu/cephtest/clone.client.2 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.2 && cd /home/ubuntu/cephtest/clone.client.2 && git checkout 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b 2026-03-10T05:36:50.145 INFO:teuthology.orchestra.run.vm05.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.2'... 2026-03-10T05:36:51.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:51 vm05 ceph-mon[50927]: pgmap v278: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:51.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:36:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:36:51.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:51 vm01 ceph-mon[47941]: pgmap v278: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:51.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:51 vm02 ceph-mon[50473]: pgmap v278: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:36:52.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:36:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:36:52.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:52 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:52.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:52 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:52.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:52 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:53 vm05 ceph-mon[50927]: pgmap v279: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:53.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:53 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:53.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:53 vm01 ceph-mon[47941]: pgmap v279: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:53.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:53 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:53.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:53 vm02 ceph-mon[50473]: pgmap v279: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:53.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:53 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:36:55.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:55 vm05 ceph-mon[50927]: pgmap v280: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:55.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:55 vm01 ceph-mon[47941]: pgmap v280: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:55.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:55 vm02 ceph-mon[50473]: pgmap v280: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:36:57.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:57 vm05 ceph-mon[50927]: pgmap v281: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:36:57.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:57 vm01 ceph-mon[47941]: pgmap v281: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:36:57.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:57 vm02 ceph-mon[50473]: pgmap v281: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:36:59.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:36:59 vm05 ceph-mon[50927]: pgmap v282: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:36:59.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:36:59 vm01 ceph-mon[47941]: pgmap v282: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:36:59.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:36:59 vm02 ceph-mon[50473]: pgmap v282: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:01.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:01 vm05 ceph-mon[50927]: pgmap v283: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:37:01.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:01 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:37:01.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:01 vm01 ceph-mon[47941]: pgmap v283: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:37:01.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:01 vm02 ceph-mon[50473]: pgmap v283: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:37:02.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:02 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:37:02.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:02 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:02.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:02 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:02.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:02 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:03 vm05 ceph-mon[50927]: pgmap v284: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:03.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:03 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:03.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:03 vm01 ceph-mon[47941]: pgmap v284: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:03.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:03 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:03.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:03 vm02 ceph-mon[50473]: pgmap v284: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:03.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:03 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:05.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:04 vm05 ceph-mon[50927]: pgmap v285: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:05.087 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:04 vm01 ceph-mon[47941]: pgmap v285: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:05.147 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:04 vm02 ceph-mon[50473]: pgmap v285: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:07.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:06 vm05 ceph-mon[50927]: pgmap v286: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:37:07.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:06 vm01 ceph-mon[47941]: pgmap v286: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:37:07.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:06 vm02 ceph-mon[50473]: pgmap v286: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:37:09.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:08 vm05 ceph-mon[50927]: pgmap v287: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:09.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:08 vm01 ceph-mon[47941]: pgmap v287: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:09.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:08 vm02 ceph-mon[50473]: pgmap v287: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:11.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:10 vm05 ceph-mon[50927]: pgmap v288: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:11.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:10 vm01 ceph-mon[47941]: pgmap v288: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:11.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:10 vm02 ceph-mon[50473]: pgmap v288: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:11.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:11 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:37:12.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:11 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:12.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:11 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:12.397 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:11 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:12.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:12 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:37:13.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:12 vm05 ceph-mon[50927]: pgmap v289: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:13.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:12 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:13.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:12 vm01 ceph-mon[47941]: pgmap v289: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:13.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:12 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:12 vm02 ceph-mon[50473]: pgmap v289: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:13.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:12 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:15.306 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:14 vm05 ceph-mon[50927]: pgmap v290: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:15.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:14 vm01 ceph-mon[47941]: pgmap v290: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:15.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:14 vm02 ceph-mon[50473]: pgmap v290: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:16.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:37:16.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:37:16.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:37:16.305 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:15 vm05 ceph-mon[50927]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:37:16.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:37:16.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:37:16.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:37:16.337 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:15 vm01 ceph-mon[47941]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:37:16.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T05:37:16.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T05:37:16.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T05:37:16.398 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:15 vm02 ceph-mon[50473]: from='mgr.14150 192.168.123.101:0/3727816590' entity='mgr.x' 2026-03-10T05:37:17.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:17 vm02 ceph-mon[50473]: pgmap v291: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:17.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:17 vm05 ceph-mon[50927]: pgmap v291: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:17.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:17 vm01 ceph-mon[47941]: pgmap v291: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:19.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:19 vm02 ceph-mon[50473]: pgmap v292: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:19.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:19 vm05 ceph-mon[50927]: pgmap v292: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:19.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:19 vm01 ceph-mon[47941]: pgmap v292: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:21.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:21 vm02 ceph-mon[50473]: pgmap v293: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:21.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:21 vm05 ceph-mon[50927]: pgmap v293: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:21.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:21 vm01 ceph-mon[47941]: pgmap v293: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:21.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:21 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:37:22.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:22 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:22.805 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:22 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:37:22.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:22 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:22.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:22 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:23.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:23 vm05 ceph-mon[50927]: pgmap v294: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:23.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:23 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:23.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:23 vm01 ceph-mon[47941]: pgmap v294: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:23.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:23 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:23.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:23 vm02 ceph-mon[50473]: pgmap v294: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:23.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:23 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:25.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:25 vm05 ceph-mon[50927]: pgmap v295: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:25.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:25 vm01 ceph-mon[47941]: pgmap v295: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:25.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:25 vm02 ceph-mon[50473]: pgmap v295: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:27.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:27 vm05 ceph-mon[50927]: pgmap v296: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:27.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:27 vm01 ceph-mon[47941]: pgmap v296: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:27.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:27 vm02 ceph-mon[50473]: pgmap v296: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:29.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:29 vm05 ceph-mon[50927]: pgmap v297: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:29.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:29 vm01 ceph-mon[47941]: pgmap v297: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:29.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:29 vm02 ceph-mon[50473]: pgmap v297: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:31.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:30 vm05 ceph-mon[50927]: pgmap v298: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:31.087 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:30 vm01 ceph-mon[47941]: pgmap v298: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:31.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:30 vm02 ceph-mon[50473]: pgmap v298: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:31.837 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:31 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:37:31.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:31 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:32.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:31 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:32.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:31 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:33.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:32 vm05 ceph-mon[50927]: pgmap v299: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:33.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:32 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:33.055 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:32 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:37:33.087 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:32 vm01 ceph-mon[47941]: pgmap v299: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:33.087 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:32 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:33.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:32 vm02 ceph-mon[50473]: pgmap v299: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:33.148 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:32 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:35.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:35 vm01 ceph-mon[47941]: pgmap v300: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:35.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:35 vm02 ceph-mon[50473]: pgmap v300: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:36.055 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:35 vm05 ceph-mon[50927]: pgmap v300: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:37.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:37 vm05 ceph-mon[50927]: pgmap v301: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:37.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:37 vm01 ceph-mon[47941]: pgmap v301: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:37.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:37 vm02 ceph-mon[50473]: pgmap v301: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:39.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:39 vm01 ceph-mon[47941]: pgmap v302: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:39.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:39 vm02 ceph-mon[50473]: pgmap v302: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:39.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:39 vm05 ceph-mon[50927]: pgmap v302: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:41.587 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:41 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:37:41.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:41 vm01 ceph-mon[47941]: pgmap v303: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:41.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:41 vm02 ceph-mon[50473]: pgmap v303: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:41.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:41 vm05 ceph-mon[50927]: pgmap v303: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-10T05:37:42.571 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:42 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:42.571 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:42 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug there is no tcmu-runner data available 2026-03-10T05:37:42.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:42 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:42.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:42 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:43.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:43 vm01 ceph-mon[47941]: pgmap v304: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:43.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:43 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:43.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:43 vm02 ceph-mon[50473]: pgmap v304: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:43.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:43 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:43.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:43 vm05 ceph-mon[50927]: pgmap v304: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:43.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:43 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:45.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:45 vm01 ceph-mon[47941]: pgmap v305: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:45.607 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:45 vm05 ceph-mon[50927]: pgmap v305: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:45.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:45 vm02 ceph-mon[50473]: pgmap v305: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-10T05:37:47.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:47 vm01 ceph-mon[47941]: pgmap v306: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:37:47.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:47 vm02 ceph-mon[50473]: pgmap v306: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:37:47.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:47 vm05 ceph-mon[50927]: pgmap v306: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:37:49.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:49 vm01 ceph-mon[47941]: pgmap v307: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:49.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:49 vm02 ceph-mon[50473]: pgmap v307: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:49.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:49 vm05 ceph-mon[50927]: pgmap v307: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr:Note: switching to '75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b'. 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr:state without impacting any branches by switching back to a branch. 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr: git switch -c 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr:Or undo this operation with: 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr: git switch - 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:37:51.625 INFO:teuthology.orchestra.run.vm05.stderr:HEAD is now at 75a68fd8ca3 qa/suites/orch/cephadm/osds: drop nvme_loop task 2026-03-10T05:37:51.630 DEBUG:teuthology.orchestra.run.vm05:> cp -- /home/ubuntu/cephtest/clone.client.2/src/test/cli-integration/rbd/rest_api_delete.t /home/ubuntu/cephtest/archive/cram.client.2 2026-03-10T05:37:51.647 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:51 vm02 ceph-mon[50473]: pgmap v308: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:37:51.687 INFO:tasks.cram:Running tests for client.0... 2026-03-10T05:37:51.687 DEBUG:teuthology.orchestra.run.vm01:> CEPH_REF=master CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.0/*.t 2026-03-10T05:37:51.733 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-mon[47941]: pgmap v308: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:37:51.733 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug there is no tcmu-runner data available 2026-03-10T05:37:51.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:51 vm05 ceph-mon[50927]: pgmap v308: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 2 op/s 2026-03-10T05:37:52.014 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:37:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.014 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:37:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.014 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (LUN.allocate) created datapool/block1 successfully 2026-03-10T05:37:52.014 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block1' to LIO backstore user:rbd 2026-03-10T05:37:52.014 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (LUN.add_dev_to_lio) Successfully added datapool/block1 to LIO 2026-03-10T05:37:52.014 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (LUN.allocate) added 'datapool/block1' to LIO and config object 2026-03-10T05:37:52.014 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:37:51] "PUT /api/_disk/datapool/block1 HTTP/1.1" 200 - 2026-03-10T05:37:52.014 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:37:51] "PUT /api/_disk/datapool/block1 HTTP/1.1" 200 - 2026-03-10T05:37:52.014 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _disk update on localhost, successful 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:127.0.0.1 - - [10/Mar/2026 05:37:51] "PUT /api/disk/datapool/block1 HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:127.0.0.1 - - [10/Mar/2026 05:37:51] "PUT /api/disk/datapool/block1 HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Gateway.create_target) created an iscsi target with iqn of 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:127.0.0.1 - - [10/Mar/2026 05:37:51] "PUT /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:127.0.0.1 - - [10/Mar/2026 05:37:51] "PUT /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug gateway validation needed for vm01.local 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:37:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:37:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug An additional 1 tpg's are required 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Gateway.create_tpg) created TPG '1' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug iqn.2003-01.com.redhat.iscsi-gw:ceph-gw - Adding the IP to the enabled tpg, allowing iSCSI logins 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:37:51] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:37:51] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _gateway update on localhost, successful 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:127.0.0.1 - - [10/Mar/2026 05:37:51] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:127.0.0.1 - - [10/Mar/2026 05:37:51] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm01.local HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug gateway validation needed for vm05.local 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:51 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.015 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _gateway update on vm05.local, successful 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:51 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:51 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:51 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:51 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:51 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:51 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:51 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:51 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:51 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Gateway.create_tpg) created TPG '1' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Gateway.create_tpg) created TPG '2' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Gateway.create_target) created an iscsi target with iqn of 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug iqn.2003-01.com.redhat.iscsi-gw:ceph-gw - Adding the IP to the enabled tpg, allowing iSCSI logins 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.292 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.337 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-10T05:37:52.337 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-10T05:37:52.337 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug An additional 1 tpg's are required 2026-03-10T05:37:52.337 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Gateway.create_tpg) created TPG '2' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-10T05:37:52.337 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-10T05:37:52.337 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-10T05:37:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _gateway update on vm01.local, successful 2026-03-10T05:37:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:127.0.0.1 - - [10/Mar/2026 05:37:52] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-10T05:37:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:127.0.0.1 - - [10/Mar/2026 05:37:52] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm05.local HTTP/1.1" 200 - 2026-03-10T05:37:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block1' to LIO backstore user:rbd 2026-03-10T05:37:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (LUN.add_dev_to_lio) Successfully added datapool/block1 to LIO 2026-03-10T05:37:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-10T05:37:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Setup group ano2 for datapool.block1 on tpg 2 (state 1, owner False, failover type 1) 2026-03-10T05:37:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Setup group ao for datapool.block1 on tpg 1 (state 0, owner True, failover type 1) 2026-03-10T05:37:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:37:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:37:52.338 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _targetlun update on vm01.local, successful 2026-03-10T05:37:52.504 INFO:tasks.cram.client.0.vm01.stdout:/home/ubuntu/cephtest/archive/cram.client.0/rest_api_create.t: passed 2026-03-10T05:37:52.504 INFO:tasks.cram.client.0.vm01.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-10T05:37:52.508 INFO:tasks.cram:Running tests for client.1... 2026-03-10T05:37:52.508 DEBUG:teuthology.orchestra.run.vm02:> CEPH_REF=master CEPH_ID="1" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.1/*.t 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block1' to LIO backstore user:rbd 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (LUN.add_dev_to_lio) Successfully added datapool/block1 to LIO 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (LUN.allocate) added block1 to LIO using wwn 'dcbb89c1-c08d-46de-9adf-298f6b142834' defined by vm01.local 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug Setup group ano2 for datapool.block1 on tpg 2 (state 1, owner False, failover type 1) 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug Setup group ao for datapool.block1 on tpg 1 (state 0, owner True, failover type 1) 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Client.define_client) iqn.1994-05.com.redhat:client added successfully 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Client.add_lun) added image 'datapool/block1' to iqn.1994-05.com.redhat:client 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:37:52.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:37:52 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.101 - - [10/Mar/2026 05:37:52] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2774973957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2774973957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2774973957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/2774973957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/4243603191"}]: dispatch 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/4243603191"}]: dispatch 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4045237870' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4045237870' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4045237870' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.101:0/4045237870' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/2774973957"}]: dispatch 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/2774973957"}]: dispatch 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1442393832' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1442393832' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1442393832' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:37:52.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:52 vm05 ceph-mon[50927]: from='client.? 192.168.123.105:0/1442393832' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.105:0/2483483664"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2774973957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2774973957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2774973957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/2774973957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/4243603191"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/4243603191"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4045237870' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4045237870' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4045237870' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.101:0/4045237870' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/2774973957"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/2774973957"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1442393832' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1442393832' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1442393832' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:37:52.636 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:52 vm02 ceph-mon[50473]: from='client.? 192.168.123.105:0/1442393832' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.105:0/2483483664"}]: dispatch 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _targetlun update on vm05.local, successful 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:127.0.0.1 - - [10/Mar/2026 05:37:52] "PUT /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:127.0.0.1 - - [10/Mar/2026 05:37:52] "PUT /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:37:52] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:37:52] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:37:52] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:37:52] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Client.define_client) iqn.1994-05.com.redhat:client added successfully 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:37:52] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:37:52] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _client update on localhost, successful 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _client update on vm05.local, successful 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:127.0.0.1 - - [10/Mar/2026 05:37:52] "PUT /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:127.0.0.1 - - [10/Mar/2026 05:37:52] "PUT /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:37:52] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:37:52] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:37:52] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:37:52] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Client.add_lun) added image 'datapool/block1' to iqn.1994-05.com.redhat:client 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::1 - - [10/Mar/2026 05:37:52] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::1 - - [10/Mar/2026 05:37:52] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _clientlun update on localhost, successful 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug _clientlun update on vm05.local, successful 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:127.0.0.1 - - [10/Mar/2026 05:37:52] "PUT /api/clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:127.0.0.1 - - [10/Mar/2026 05:37:52] "PUT /api/clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:37:52.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:52.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2774973957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:37:52.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2774973957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:37:52.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2774973957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:37:52.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/2774973957' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/4243603191"}]: dispatch 2026-03-10T05:37:52.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/4243603191"}]: dispatch 2026-03-10T05:37:52.838 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4045237870' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:37:52.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4045237870' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:37:52.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4045237870' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:37:52.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.101:0/4045237870' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/2774973957"}]: dispatch 2026-03-10T05:37:52.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.? ' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.101:0/2774973957"}]: dispatch 2026-03-10T05:37:52.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1442393832' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"}]: dispatch 2026-03-10T05:37:52.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1442393832' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"}]: dispatch 2026-03-10T05:37:52.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1442393832' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"}]: dispatch 2026-03-10T05:37:52.839 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:52 vm01 ceph-mon[47941]: from='client.? 192.168.123.105:0/1442393832' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.105:0/2483483664"}]: dispatch 2026-03-10T05:37:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:53 vm05 ceph-mon[50927]: pgmap v309: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:53.555 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:53 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:53 vm02 ceph-mon[50473]: pgmap v309: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:53.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:53 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:53.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:53 vm01 ceph-mon[47941]: pgmap v309: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:53.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:53 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:37:55.648 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:55 vm02 ceph-mon[50473]: pgmap v310: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:55.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:55 vm05 ceph-mon[50927]: pgmap v310: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:55.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:55 vm01 ceph-mon[47941]: pgmap v310: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s rd, 1 op/s 2026-03-10T05:37:57.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:57 vm05 ceph-mon[50927]: pgmap v311: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-10T05:37:57.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:57 vm01 ceph-mon[47941]: pgmap v311: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-10T05:37:57.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:57 vm02 ceph-mon[50473]: pgmap v311: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-10T05:37:59.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:37:59 vm05 ceph-mon[50927]: pgmap v312: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-10T05:37:59.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:37:59 vm01 ceph-mon[47941]: pgmap v312: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-10T05:37:59.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:37:59 vm02 ceph-mon[50473]: pgmap v312: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-10T05:38:01.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:01 vm05 ceph-mon[50927]: pgmap v313: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-10T05:38:01.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:38:01 vm01 ceph-mon[47941]: pgmap v313: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-10T05:38:01.897 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:01 vm02 ceph-mon[50473]: pgmap v313: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-10T05:38:02.762 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:02 vm02 ceph-mon[50473]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:38:02.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:02 vm05 ceph-mon[50927]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:38:02.837 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:38:02 vm01 ceph-mon[47941]: from='client.14445 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:38:02.933 INFO:tasks.cram.client.1.vm02.stdout:/home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t: passed 2026-03-10T05:38:02.933 INFO:tasks.cram.client.1.vm02.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-10T05:38:02.939 INFO:tasks.cram:Running tests for client.2... 2026-03-10T05:38:02.939 DEBUG:teuthology.orchestra.run.vm05:> CEPH_REF=master CEPH_ID="2" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.2/*.t 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:38:03] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:38:03] "GET /api/config HTTP/1.1" 200 - 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:38:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:38:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug _client update on vm01.local, successful 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Client.delete) deleted NodeACL for iqn.1994-05.com.redhat:client 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::1 - - [10/Mar/2026 05:38:03] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::1 - - [10/Mar/2026 05:38:03] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug _client update on localhost, successful 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:127.0.0.1 - - [10/Mar/2026 05:38:03] "DELETE /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:127.0.0.1 - - [10/Mar/2026 05:38:03] "DELETE /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug _targetlun update on vm01.local, successful 2026-03-10T05:38:03.245 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug LUN unmap request received, config commit to be performed by vm05.local 2026-03-10T05:38:03.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:38:03.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:38:03.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Client.delete) deleted NodeACL for iqn.1994-05.com.redhat:client 2026-03-10T05:38:03.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:38:03.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-10T05:38:03.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:38:03.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:38:03.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug LUN unmap request received, config commit to be performed by vm05.local 2026-03-10T05:38:03.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:38:03.288 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:38:03.555 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug _targetlun update on vm05.local, successful 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:127.0.0.1 - - [10/Mar/2026 05:38:03] "DELETE /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:127.0.0.1 - - [10/Mar/2026 05:38:03] "DELETE /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug _target update on vm01.local, successful 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug Removing target configuration 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug _target update on vm05.local, successful 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:127.0.0.1 - - [10/Mar/2026 05:38:03] "DELETE /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:127.0.0.1 - - [10/Mar/2026 05:38:03] "DELETE /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:38:03.556 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug LUN deletion request received, rbd removal to be performed by vm05.local 2026-03-10T05:38:03.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:03 vm05 ceph-mon[50927]: pgmap v314: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-10T05:38:03.556 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:03 vm05 ceph-mon[50927]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:38:03.561 INFO:tasks.cram.client.2.vm05.stdout:/home/ubuntu/cephtest/archive/cram.client.2/rest_api_delete.t: passed 2026-03-10T05:38:03.561 INFO:tasks.cram.client.2.vm05.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-10T05:38:03.566 DEBUG:teuthology.orchestra.run.vm01:> test -f /home/ubuntu/cephtest/archive/cram.client.0/rest_api_create.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.0/rest_api_create.t 2026-03-10T05:38:03.582 DEBUG:teuthology.orchestra.run.vm01:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.0 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.0 2026-03-10T05:38:03.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-mon[47941]: pgmap v314: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-10T05:38:03.587 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-mon[47941]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:38:03.587 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:38:03.587 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "GET /api/_ping HTTP/1.1" 200 - 2026-03-10T05:38:03.587 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-10T05:38:03.587 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Removing target configuration 2026-03-10T05:38:03.587 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:38:03.587 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-10T05:38:03.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:03 vm02 ceph-mon[50473]: pgmap v314: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-10T05:38:03.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:03 vm02 ceph-mon[50473]: from='client.24364 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-10T05:38:04.001 DEBUG:teuthology.orchestra.run.vm02:> test -f /home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t 2026-03-10T05:38:04.017 DEBUG:teuthology.orchestra.run.vm02:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.1 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.1 2026-03-10T05:38:04.055 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "DELETE /api/_disk/datapool/block1 HTTP/1.1" 200 - 2026-03-10T05:38:04.055 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:192.168.123.105 - - [10/Mar/2026 05:38:03] "DELETE /api/_disk/datapool/block1 HTTP/1.1" 200 - 2026-03-10T05:38:04.055 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug _disk update on vm05.local, successful 2026-03-10T05:38:04.055 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug ::ffff:127.0.0.1 - - [10/Mar/2026 05:38:03] "DELETE /api/disk/datapool/block1 HTTP/1.1" 200 - 2026-03-10T05:38:04.055 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:03 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: ::ffff:127.0.0.1 - - [10/Mar/2026 05:38:03] "DELETE /api/disk/datapool/block1 HTTP/1.1" 200 - 2026-03-10T05:38:04.087 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:03 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Change detected - internal 24 / xattr 26 refreshing 2026-03-10T05:38:04.479 DEBUG:teuthology.orchestra.run.vm05:> test -f /home/ubuntu/cephtest/archive/cram.client.2/rest_api_delete.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.2/rest_api_delete.t 2026-03-10T05:38:04.497 DEBUG:teuthology.orchestra.run.vm05:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.2 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.2 2026-03-10T05:38:04.886 DEBUG:teuthology.run_tasks:Unwinding manager ceph_iscsi_client 2026-03-10T05:38:04.888 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-10T05:38:04.890 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-10T05:38:04.890 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-10T05:38:04.891 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-10T05:38:04.893 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-10T05:38:04.929 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-10T05:38:04.929 DEBUG:teuthology.orchestra.run.vm01:> 2026-03-10T05:38:04.929 DEBUG:teuthology.orchestra.run.vm01:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-10T05:38:04.929 DEBUG:teuthology.orchestra.run.vm01:> sudo yum -y remove $d || true 2026-03-10T05:38:04.929 DEBUG:teuthology.orchestra.run.vm01:> done 2026-03-10T05:38:04.935 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-10T05:38:04.936 DEBUG:teuthology.orchestra.run.vm02:> 2026-03-10T05:38:04.936 DEBUG:teuthology.orchestra.run.vm02:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-10T05:38:04.936 DEBUG:teuthology.orchestra.run.vm02:> sudo yum -y remove $d || true 2026-03-10T05:38:04.936 DEBUG:teuthology.orchestra.run.vm02:> done 2026-03-10T05:38:04.940 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-10T05:38:04.940 DEBUG:teuthology.orchestra.run.vm05:> 2026-03-10T05:38:04.941 DEBUG:teuthology.orchestra.run.vm05:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-10T05:38:04.941 DEBUG:teuthology.orchestra.run.vm05:> sudo yum -y remove $d || true 2026-03-10T05:38:04.941 DEBUG:teuthology.orchestra.run.vm05:> done 2026-03-10T05:38:05.161 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:05.162 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:05.162 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-10T05:38:05.162 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:05.162 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-10T05:38:05.162 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-10T05:38:05.162 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-10T05:38:05.162 INFO:teuthology.orchestra.run.vm01.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-10T05:38:05.162 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:05.162 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-10T05:38:05.162 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:05.162 INFO:teuthology.orchestra.run.vm01.stdout:Remove 2 Packages 2026-03-10T05:38:05.162 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:05.162 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 39 M 2026-03-10T05:38:05.162 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-10T05:38:05.164 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-10T05:38:05.164 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-10T05:38:05.165 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:05.166 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:05.166 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-10T05:38:05.166 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:05.166 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-10T05:38:05.166 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-10T05:38:05.166 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-10T05:38:05.166 INFO:teuthology.orchestra.run.vm02.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-10T05:38:05.166 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:05.166 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-10T05:38:05.166 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:05.166 INFO:teuthology.orchestra.run.vm02.stdout:Remove 2 Packages 2026-03-10T05:38:05.166 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:05.166 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 39 M 2026-03-10T05:38:05.166 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-10T05:38:05.168 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-10T05:38:05.168 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-10T05:38:05.175 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:05.176 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:05.176 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-10T05:38:05.176 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:05.176 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-10T05:38:05.176 INFO:teuthology.orchestra.run.vm05.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-10T05:38:05.176 INFO:teuthology.orchestra.run.vm05.stdout:Removing unused dependencies: 2026-03-10T05:38:05.176 INFO:teuthology.orchestra.run.vm05.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-10T05:38:05.176 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:05.176 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-10T05:38:05.176 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:05.176 INFO:teuthology.orchestra.run.vm05.stdout:Remove 2 Packages 2026-03-10T05:38:05.176 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:05.176 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 39 M 2026-03-10T05:38:05.176 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-10T05:38:05.179 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-10T05:38:05.179 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-10T05:38:05.179 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-10T05:38:05.179 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-10T05:38:05.183 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-10T05:38:05.183 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-10T05:38:05.194 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-10T05:38:05.194 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-10T05:38:05.212 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-10T05:38:05.215 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-10T05:38:05.227 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-10T05:38:05.241 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:05.241 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:05.241 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T05:38:05.241 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-10T05:38:05.241 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-10T05:38:05.241 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:05.241 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:05.242 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:05.242 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T05:38:05.242 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-10T05:38:05.242 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-10T05:38:05.242 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:05.244 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:05.245 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:05.253 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:05.254 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:05.255 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:05.255 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T05:38:05.255 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-10T05:38:05.255 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-10T05:38:05.255 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:05.256 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:05.258 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:05.269 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:05.269 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T05:38:05.271 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T05:38:05.286 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T05:38:05.371 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T05:38:05.371 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:05.375 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T05:38:05.375 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:05.387 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T05:38:05.387 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:05.439 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T05:38:05.439 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:05.439 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-10T05:38:05.439 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-10T05:38:05.439 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:05.439 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:05.440 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T05:38:05.440 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:05.440 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-10T05:38:05.440 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-10T05:38:05.440 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:05.440 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:05.450 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T05:38:05.450 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:05.450 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-10T05:38:05.450 INFO:teuthology.orchestra.run.vm05.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-10T05:38:05.450 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:05.450 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:05.502 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:05 vm02 ceph-mon[50473]: pgmap v315: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-10T05:38:05.648 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout:Remove 4 Packages 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 212 M 2026-03-10T05:38:05.649 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-10T05:38:05.652 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-10T05:38:05.652 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout:Remove 4 Packages 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:05.667 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 212 M 2026-03-10T05:38:05.668 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-10T05:38:05.670 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-10T05:38:05.670 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-10T05:38:05.672 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:05.673 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:05.673 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-10T05:38:05.673 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:05.673 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-10T05:38:05.673 INFO:teuthology.orchestra.run.vm05.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-10T05:38:05.673 INFO:teuthology.orchestra.run.vm05.stdout:Removing unused dependencies: 2026-03-10T05:38:05.673 INFO:teuthology.orchestra.run.vm05.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-10T05:38:05.673 INFO:teuthology.orchestra.run.vm05.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-10T05:38:05.673 INFO:teuthology.orchestra.run.vm05.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-10T05:38:05.673 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:05.673 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-10T05:38:05.673 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:05.673 INFO:teuthology.orchestra.run.vm05.stdout:Remove 4 Packages 2026-03-10T05:38:05.673 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:05.674 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 212 M 2026-03-10T05:38:05.674 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-10T05:38:05.676 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-10T05:38:05.676 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-10T05:38:05.677 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-10T05:38:05.677 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-10T05:38:05.693 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-10T05:38:05.693 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-10T05:38:05.704 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-10T05:38:05.705 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-10T05:38:05.746 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-10T05:38:05.753 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T05:38:05.756 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-10T05:38:05.759 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-10T05:38:05.776 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T05:38:05.780 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-10T05:38:05.787 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T05:38:05.789 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-10T05:38:05.789 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-10T05:38:05.793 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-10T05:38:05.796 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T05:38:05.798 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-10T05:38:05.802 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-10T05:38:05.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:05 vm05 ceph-mon[50927]: pgmap v315: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-10T05:38:05.809 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T05:38:05.822 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T05:38:05.831 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:38:05 vm01 ceph-mon[47941]: pgmap v315: 4 pgs: 4 active+clean; 452 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-10T05:38:05.856 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T05:38:05.856 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T05:38:05.856 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-10T05:38:05.856 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-10T05:38:05.889 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T05:38:05.889 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T05:38:05.889 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-10T05:38:05.889 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-10T05:38:05.904 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T05:38:05.904 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T05:38:05.904 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-10T05:38:05.904 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-10T05:38:05.924 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-10T05:38:05.924 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:05.924 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-10T05:38:05.924 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-10T05:38:05.924 INFO:teuthology.orchestra.run.vm02.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T05:38:05.924 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:05.924 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:05.953 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-10T05:38:05.953 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:05.953 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-10T05:38:05.953 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-10T05:38:05.953 INFO:teuthology.orchestra.run.vm01.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T05:38:05.953 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:05.953 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:05.955 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-10T05:38:05.955 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:05.955 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-10T05:38:05.955 INFO:teuthology.orchestra.run.vm05.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-10T05:38:05.955 INFO:teuthology.orchestra.run.vm05.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T05:38:05.955 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:05.955 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:06.153 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout:Remove 8 Packages 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 28 M 2026-03-10T05:38:06.154 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-10T05:38:06.157 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-10T05:38:06.157 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-10T05:38:06.159 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:06.160 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:06.160 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-10T05:38:06.160 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:06.160 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-10T05:38:06.160 INFO:teuthology.orchestra.run.vm05.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-10T05:38:06.160 INFO:teuthology.orchestra.run.vm05.stdout:Removing unused dependencies: 2026-03-10T05:38:06.161 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-10T05:38:06.161 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-10T05:38:06.161 INFO:teuthology.orchestra.run.vm05.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-10T05:38:06.161 INFO:teuthology.orchestra.run.vm05.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-10T05:38:06.161 INFO:teuthology.orchestra.run.vm05.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-10T05:38:06.161 INFO:teuthology.orchestra.run.vm05.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-10T05:38:06.161 INFO:teuthology.orchestra.run.vm05.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-10T05:38:06.161 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:06.161 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-10T05:38:06.161 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:06.161 INFO:teuthology.orchestra.run.vm05.stdout:Remove 8 Packages 2026-03-10T05:38:06.161 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:06.161 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 28 M 2026-03-10T05:38:06.161 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-10T05:38:06.164 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-10T05:38:06.164 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-10T05:38:06.181 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-10T05:38:06.182 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-10T05:38:06.188 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-10T05:38:06.188 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-10T05:38:06.193 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout:Remove 8 Packages 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 28 M 2026-03-10T05:38:06.194 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-10T05:38:06.197 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-10T05:38:06.197 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-10T05:38:06.223 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-10T05:38:06.223 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-10T05:38:06.225 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-10T05:38:06.230 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T05:38:06.232 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-10T05:38:06.234 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-10T05:38:06.235 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-10T05:38:06.237 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T05:38:06.238 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-10T05:38:06.240 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-10T05:38:06.241 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-10T05:38:06.242 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-10T05:38:06.244 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-10T05:38:06.245 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-10T05:38:06.248 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-10T05:38:06.250 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-10T05:38:06.265 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T05:38:06.265 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:06.265 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T05:38:06.265 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-10T05:38:06.265 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-10T05:38:06.265 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:06.266 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T05:38:06.266 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-10T05:38:06.272 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T05:38:06.275 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T05:38:06.275 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-10T05:38:06.276 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T05:38:06.276 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:06.276 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T05:38:06.276 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-10T05:38:06.276 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-10T05:38:06.276 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:06.276 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T05:38:06.277 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-10T05:38:06.280 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-10T05:38:06.283 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T05:38:06.283 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-10T05:38:06.285 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-10T05:38:06.297 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T05:38:06.297 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:06.297 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T05:38:06.297 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-10T05:38:06.297 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-10T05:38:06.297 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:06.298 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T05:38:06.310 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T05:38:06.310 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:06.310 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T05:38:06.310 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-10T05:38:06.310 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-10T05:38:06.310 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:06.311 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T05:38:06.311 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:06.311 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T05:38:06.311 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-10T05:38:06.311 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-10T05:38:06.311 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:06.311 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T05:38:06.312 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T05:38:06.321 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T05:38:06.349 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T05:38:06.349 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:06.349 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T05:38:06.350 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-10T05:38:06.350 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-10T05:38:06.350 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:06.351 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T05:38:06.392 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T05:38:06.392 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T05:38:06.392 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-10T05:38:06.392 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-10T05:38:06.392 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-10T05:38:06.392 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-10T05:38:06.392 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-10T05:38:06.392 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-10T05:38:06.400 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T05:38:06.400 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T05:38:06.400 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-10T05:38:06.400 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-10T05:38:06.400 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-10T05:38:06.400 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-10T05:38:06.400 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-10T05:38:06.400 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-10T05:38:06.448 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T05:38:06.448 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T05:38:06.448 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-10T05:38:06.448 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-10T05:38:06.448 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-10T05:38:06.448 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-10T05:38:06.448 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-10T05:38:06.448 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-10T05:38:06.455 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-10T05:38:06.455 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:06.455 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-10T05:38:06.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:06.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:06.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:06.455 INFO:teuthology.orchestra.run.vm05.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T05:38:06.455 INFO:teuthology.orchestra.run.vm05.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T05:38:06.455 INFO:teuthology.orchestra.run.vm05.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T05:38:06.455 INFO:teuthology.orchestra.run.vm05.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T05:38:06.455 INFO:teuthology.orchestra.run.vm05.stdout: zip-3.0-35.el9.x86_64 2026-03-10T05:38:06.455 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:06.455 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:06.459 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-10T05:38:06.459 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:06.459 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-10T05:38:06.459 INFO:teuthology.orchestra.run.vm02.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:06.459 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:06.459 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:06.459 INFO:teuthology.orchestra.run.vm02.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T05:38:06.459 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T05:38:06.459 INFO:teuthology.orchestra.run.vm02.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T05:38:06.459 INFO:teuthology.orchestra.run.vm02.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T05:38:06.459 INFO:teuthology.orchestra.run.vm02.stdout: zip-3.0-35.el9.x86_64 2026-03-10T05:38:06.459 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:06.459 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:06.511 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-10T05:38:06.511 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:06.511 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-10T05:38:06.511 INFO:teuthology.orchestra.run.vm01.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:06.511 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:06.511 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:06.511 INFO:teuthology.orchestra.run.vm01.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T05:38:06.511 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T05:38:06.511 INFO:teuthology.orchestra.run.vm01.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T05:38:06.511 INFO:teuthology.orchestra.run.vm01.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T05:38:06.511 INFO:teuthology.orchestra.run.vm01.stdout: zip-3.0-35.el9.x86_64 2026-03-10T05:38:06.511 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:06.511 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:06.677 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:06.683 INFO:teuthology.orchestra.run.vm02.stdout:=========================================================================================== 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout:=========================================================================================== 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout:Removing dependent packages: 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-10T05:38:06.684 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-10T05:38:06.685 INFO:teuthology.orchestra.run.vm02.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-10T05:38:06.686 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-10T05:38:06.686 INFO:teuthology.orchestra.run.vm02.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-10T05:38:06.686 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-10T05:38:06.686 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-10T05:38:06.686 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:06.686 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-10T05:38:06.686 INFO:teuthology.orchestra.run.vm02.stdout:=========================================================================================== 2026-03-10T05:38:06.686 INFO:teuthology.orchestra.run.vm02.stdout:Remove 102 Packages 2026-03-10T05:38:06.686 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:06.686 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 613 M 2026-03-10T05:38:06.686 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-10T05:38:06.690 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout:=========================================================================================== 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout:=========================================================================================== 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout:Removing dependent packages: 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout:Removing unused dependencies: 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-10T05:38:06.697 INFO:teuthology.orchestra.run.vm05.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-10T05:38:06.698 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout:=========================================================================================== 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout:Remove 102 Packages 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 613 M 2026-03-10T05:38:06.699 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-10T05:38:06.713 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-10T05:38:06.714 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-10T05:38:06.728 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-10T05:38:06.728 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-10T05:38:06.737 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout:=========================================================================================== 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout:=========================================================================================== 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-10T05:38:06.743 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-10T05:38:06.744 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout:=========================================================================================== 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout:Remove 102 Packages 2026-03-10T05:38:06.745 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:06.746 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 613 M 2026-03-10T05:38:06.746 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-10T05:38:06.773 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-10T05:38:06.773 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-10T05:38:06.830 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-10T05:38:06.830 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-10T05:38:06.841 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-10T05:38:06.842 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-10T05:38:06.890 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-10T05:38:06.890 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-10T05:38:06.988 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-10T05:38:06.989 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-10T05:38:06.996 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-10T05:38:06.997 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-10T05:38:06.997 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-10T05:38:07.006 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-10T05:38:07.029 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T05:38:07.029 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.029 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T05:38:07.029 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-10T05:38:07.029 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-10T05:38:07.029 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:07.029 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T05:38:07.029 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T05:38:07.029 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.029 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T05:38:07.029 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-10T05:38:07.029 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-10T05:38:07.030 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:07.030 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T05:38:07.043 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T05:38:07.044 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T05:38:07.046 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-10T05:38:07.046 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-10T05:38:07.055 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-10T05:38:07.069 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-10T05:38:07.070 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-10T05:38:07.070 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-10T05:38:07.070 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-10T05:38:07.081 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T05:38:07.081 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.081 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T05:38:07.081 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-10T05:38:07.081 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-10T05:38:07.081 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:07.082 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T05:38:07.096 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T05:38:07.122 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-10T05:38:07.122 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-10T05:38:07.132 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-10T05:38:07.138 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-10T05:38:07.143 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-10T05:38:07.147 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-10T05:38:07.148 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T05:38:07.148 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-10T05:38:07.153 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-10T05:38:07.153 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T05:38:07.160 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T05:38:07.167 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T05:38:07.168 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-10T05:38:07.172 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-10T05:38:07.175 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-10T05:38:07.179 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-10T05:38:07.181 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-10T05:38:07.183 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-10T05:38:07.187 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-10T05:38:07.188 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-10T05:38:07.193 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-10T05:38:07.194 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-10T05:38:07.200 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-10T05:38:07.200 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T05:38:07.214 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T05:38:07.214 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.214 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T05:38:07.214 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-10T05:38:07.214 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-10T05:38:07.215 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:07.215 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T05:38:07.218 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T05:38:07.218 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.218 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T05:38:07.218 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-10T05:38:07.218 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-10T05:38:07.218 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:07.220 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T05:38:07.222 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-10T05:38:07.223 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T05:38:07.226 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-10T05:38:07.232 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T05:38:07.233 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T05:38:07.236 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-10T05:38:07.240 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-10T05:38:07.253 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T05:38:07.253 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.253 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T05:38:07.253 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:07.253 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T05:38:07.253 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.253 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T05:38:07.253 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:07.261 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T05:38:07.262 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T05:38:07.266 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T05:38:07.266 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.266 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T05:38:07.266 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-10T05:38:07.266 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-10T05:38:07.266 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:07.272 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T05:38:07.274 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T05:38:07.276 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T05:38:07.277 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-10T05:38:07.279 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-10T05:38:07.282 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-10T05:38:07.283 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T05:38:07.286 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-10T05:38:07.287 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-10T05:38:07.291 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-10T05:38:07.297 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-10T05:38:07.301 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-10T05:38:07.303 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T05:38:07.303 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.303 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T05:38:07.303 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:07.310 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-10T05:38:07.311 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T05:38:07.313 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-10T05:38:07.317 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-10T05:38:07.320 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-10T05:38:07.322 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T05:38:07.325 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-10T05:38:07.327 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-10T05:38:07.329 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-10T05:38:07.330 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-10T05:38:07.334 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-10T05:38:07.334 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-10T05:38:07.336 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-10T05:38:07.344 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-10T05:38:07.359 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-10T05:38:07.365 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-10T05:38:07.366 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-10T05:38:07.367 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-10T05:38:07.372 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-10T05:38:07.375 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-10T05:38:07.375 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-10T05:38:07.377 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-10T05:38:07.379 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-10T05:38:07.385 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-10T05:38:07.387 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-10T05:38:07.387 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-10T05:38:07.396 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-10T05:38:07.396 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-10T05:38:07.401 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-10T05:38:07.401 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-10T05:38:07.405 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-10T05:38:07.410 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-10T05:38:07.419 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-10T05:38:07.426 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-10T05:38:07.430 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-10T05:38:07.439 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-10T05:38:07.449 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-10T05:38:07.450 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-10T05:38:07.458 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-10T05:38:07.506 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-10T05:38:07.514 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-10T05:38:07.520 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-10T05:38:07.528 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-10T05:38:07.540 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T05:38:07.540 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-10T05:38:07.540 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:07.540 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T05:38:07.543 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T05:38:07.544 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-10T05:38:07.544 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:07.544 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T05:38:07.563 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-10T05:38:07.566 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T05:38:07.572 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T05:38:07.577 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-10T05:38:07.582 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-10T05:38:07.587 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-10T05:38:07.589 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-10T05:38:07.590 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-10T05:38:07.592 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-10T05:38:07.593 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T05:38:07.593 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-10T05:38:07.593 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:07.593 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T05:38:07.595 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-10T05:38:07.598 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-10T05:38:07.600 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-10T05:38:07.618 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T05:38:07.618 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.618 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T05:38:07.618 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-10T05:38:07.618 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-10T05:38:07.618 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:07.618 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T05:38:07.621 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T05:38:07.627 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T05:38:07.628 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.628 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T05:38:07.628 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-10T05:38:07.628 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-10T05:38:07.628 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:07.628 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T05:38:07.630 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T05:38:07.634 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-10T05:38:07.636 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-10T05:38:07.638 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-10T05:38:07.639 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-10T05:38:07.640 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T05:38:07.641 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-10T05:38:07.645 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-10T05:38:07.645 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-10T05:38:07.646 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-10T05:38:07.647 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-10T05:38:07.648 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-10T05:38:07.650 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-10T05:38:07.650 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-10T05:38:07.651 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-10T05:38:07.652 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-10T05:38:07.655 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-10T05:38:07.656 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-10T05:38:07.660 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-10T05:38:07.665 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-10T05:38:07.674 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T05:38:07.674 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.674 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T05:38:07.674 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-10T05:38:07.674 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-10T05:38:07.674 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:07.676 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T05:38:07.687 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T05:38:07.691 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-10T05:38:07.694 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-10T05:38:07.696 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-10T05:38:07.699 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-10T05:38:07.702 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-10T05:38:07.706 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-10T05:38:07.706 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-10T05:38:07.711 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-10T05:38:07.713 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-10T05:38:07.718 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-10T05:38:07.722 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-10T05:38:07.726 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-10T05:38:07.727 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-10T05:38:07.728 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-10T05:38:07.729 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-10T05:38:07.732 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-10T05:38:07.733 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-10T05:38:07.735 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-10T05:38:07.735 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-10T05:38:07.739 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-10T05:38:07.744 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-10T05:38:07.758 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T05:38:07.758 INFO:teuthology.orchestra.run.vm05.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.758 INFO:teuthology.orchestra.run.vm05.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T05:38:07.758 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:07.760 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T05:38:07.763 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-10T05:38:07.768 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T05:38:07.768 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T05:38:07.768 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.768 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T05:38:07.768 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:07.769 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T05:38:07.770 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-10T05:38:07.772 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-10T05:38:07.775 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-10T05:38:07.776 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-10T05:38:07.778 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-10T05:38:07.779 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T05:38:07.779 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-10T05:38:07.780 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-10T05:38:07.781 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-10T05:38:07.783 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-10T05:38:07.784 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-10T05:38:07.785 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-10T05:38:07.786 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-10T05:38:07.787 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-10T05:38:07.787 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-10T05:38:07.788 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-10T05:38:07.790 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-10T05:38:07.791 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-10T05:38:07.792 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-10T05:38:07.793 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-10T05:38:07.795 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-10T05:38:07.796 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-10T05:38:07.799 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-10T05:38:07.800 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-10T05:38:07.805 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:07 vm05 ceph-mon[50927]: pgmap v316: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 102 KiB/s rd, 42 KiB/s wr, 74 op/s 2026-03-10T05:38:07.805 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-10T05:38:07.807 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-10T05:38:07.807 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-10T05:38:07.809 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-10T05:38:07.812 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-10T05:38:07.812 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-10T05:38:07.814 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:38:07 vm01 ceph-mon[47941]: pgmap v316: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 102 KiB/s rd, 42 KiB/s wr, 74 op/s 2026-03-10T05:38:07.814 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-10T05:38:07.816 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-10T05:38:07.818 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-10T05:38:07.819 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-10T05:38:07.821 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T05:38:07.821 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T05:38:07.821 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T05:38:07.821 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:07.821 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T05:38:07.821 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-10T05:38:07.824 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-10T05:38:07.827 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-10T05:38:07.829 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-10T05:38:07.830 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T05:38:07.831 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-10T05:38:07.832 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-10T05:38:07.834 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-10T05:38:07.834 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-10T05:38:07.837 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-10T05:38:07.837 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-10T05:38:07.838 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-10T05:38:07.840 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-10T05:38:07.841 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-10T05:38:07.843 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-10T05:38:07.844 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-10T05:38:07.844 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-10T05:38:07.846 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-10T05:38:07.847 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-10T05:38:07.847 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-10T05:38:07.848 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-10T05:38:07.851 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-10T05:38:07.851 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-10T05:38:07.853 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-10T05:38:07.854 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-10T05:38:07.857 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-10T05:38:07.859 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-10T05:38:07.859 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-10T05:38:07.860 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-10T05:38:07.865 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-10T05:38:07.865 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-10T05:38:07.867 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-10T05:38:07.868 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-10T05:38:07.869 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-10T05:38:07.870 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-10T05:38:07.873 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-10T05:38:07.875 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-10T05:38:07.876 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-10T05:38:07.878 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-10T05:38:07.878 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-10T05:38:07.880 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-10T05:38:07.882 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-10T05:38:07.882 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-10T05:38:07.883 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-10T05:38:07.885 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-10T05:38:07.887 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-10T05:38:07.888 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-10T05:38:07.888 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-10T05:38:07.889 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-10T05:38:07.892 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-10T05:38:07.893 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-10T05:38:07.895 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-10T05:38:07.898 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:07 vm02 ceph-mon[50473]: pgmap v316: 4 pgs: 4 active+clean; 449 KiB data, 216 MiB used, 160 GiB / 160 GiB avail; 102 KiB/s rd, 42 KiB/s wr, 74 op/s 2026-03-10T05:38:07.898 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-10T05:38:07.899 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-10T05:38:07.902 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-10T05:38:07.907 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-10T05:38:07.909 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-10T05:38:07.916 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-10T05:38:07.920 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-10T05:38:07.923 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-10T05:38:07.929 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T05:38:07.929 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-10T05:38:07.929 INFO:teuthology.orchestra.run.vm05.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-10T05:38:07.929 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:07.932 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-10T05:38:07.938 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T05:38:07.938 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-10T05:38:07.941 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-10T05:38:07.942 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T05:38:07.942 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-10T05:38:07.942 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-10T05:38:07.942 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:07.944 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-10T05:38:07.945 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-10T05:38:07.951 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T05:38:07.952 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-10T05:38:07.956 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-10T05:38:07.972 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T05:38:07.972 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-10T05:38:07.990 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-10T05:38:07.996 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-10T05:38:08.000 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T05:38:08.000 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-10T05:38:08.000 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-10T05:38:08.002 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T05:38:08.002 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-10T05:38:08.002 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-10T05:38:08.002 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:08.002 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-10T05:38:08.002 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-10T05:38:08.011 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T05:38:08.014 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-10T05:38:08.019 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-10T05:38:08.024 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-10T05:38:08.026 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-10T05:38:08.026 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-10T05:38:08.052 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T05:38:08.052 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-10T05:38:08.066 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:38:07 vm01 systemd[1]: Stopping Ceph osd.0 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:38:08.067 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:38:07 vm01 systemd[1]: Stopping Ceph mgr.x for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:38:08.067 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:38:07 vm01 systemd[1]: Stopping Ceph mon.a for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:38:08.067 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:38:07 vm01 systemd[1]: Stopping Ceph osd.1 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:38:08.072 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-10T05:38:08.077 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-10T05:38:08.081 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-10T05:38:08.084 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-10T05:38:08.084 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-10T05:38:08.205 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:38:07 vm02 systemd[1]: Stopping Ceph osd.2 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:38:08.206 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:38:08 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-2[53628]: 2026-03-10T05:38:08.119+0000 7f5dce8b9640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T05:38:08.206 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:38:08 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-2[53628]: 2026-03-10T05:38:08.119+0000 7f5dce8b9640 -1 osd.2 58 *** Got signal Terminated *** 2026-03-10T05:38:08.206 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:38:08 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-2[53628]: 2026-03-10T05:38:08.119+0000 7f5dce8b9640 -1 osd.2 58 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T05:38:08.206 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:38:07 vm02 systemd[1]: Stopping Ceph osd.3 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:38:08.206 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:38:08 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-3[58807]: 2026-03-10T05:38:08.121+0000 7fa5141a1640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.3 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T05:38:08.206 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:38:08 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-3[58807]: 2026-03-10T05:38:08.121+0000 7fa5141a1640 -1 osd.3 58 *** Got signal Terminated *** 2026-03-10T05:38:08.206 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:38:08 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-3[58807]: 2026-03-10T05:38:08.121+0000 7fa5141a1640 -1 osd.3 58 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T05:38:08.206 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:38:07 vm02 systemd[1]: Stopping Ceph osd.4 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:38:08.206 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:07 vm02 systemd[1]: Stopping Ceph mon.b for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:38:08.206 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:08 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-b[50469]: 2026-03-10T05:38:08.200+0000 7f8e36906640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T05:38:08.206 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:08 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-b[50469]: 2026-03-10T05:38:08.200+0000 7f8e36906640 -1 mon.b@2(peon) e3 *** Got Signal Terminated *** 2026-03-10T05:38:08.253 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:38:07 vm05 systemd[1]: Stopping Ceph osd.7 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:38:08.254 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:38:07 vm05 systemd[1]: Stopping Ceph osd.6 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:38:08.254 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 10 05:38:07 vm05 systemd[1]: Stopping Ceph osd.5 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:38:08.254 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:07 vm05 systemd[1]: Stopping Ceph mon.c for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:38:08.463 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:38:08 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-4[63840]: 2026-03-10T05:38:08.233+0000 7fb2b9bcb640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.4 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T05:38:08.463 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:38:08 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-4[63840]: 2026-03-10T05:38:08.233+0000 7fb2b9bcb640 -1 osd.4 58 *** Got signal Terminated *** 2026-03-10T05:38:08.463 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:38:08 vm02 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-4[63840]: 2026-03-10T05:38:08.233+0000 7fb2b9bcb640 -1 osd.4 58 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T05:38:08.463 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:08 vm02 podman[73640]: 2026-03-10 05:38:08.334098224 +0000 UTC m=+0.162094124 container died 4da50b6ed5ae4cef53cd890564d28e9b382aef725677cfc4793bb85babfc3824 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-b, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True) 2026-03-10T05:38:08.463 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:08 vm02 podman[73640]: 2026-03-10 05:38:08.362870873 +0000 UTC m=+0.190866773 container remove 4da50b6ed5ae4cef53cd890564d28e9b382aef725677cfc4793bb85babfc3824 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-b, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, org.label-schema.schema-version=1.0, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, org.label-schema.build-date=20260223) 2026-03-10T05:38:08.463 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:08 vm02 bash[73640]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-b 2026-03-10T05:38:08.463 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:08 vm02 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.b.service: Deactivated successfully. 2026-03-10T05:38:08.463 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:08 vm02 systemd[1]: Stopped Ceph mon.b for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:38:08.463 INFO:journalctl@ceph.mon.b.vm02.stdout:Mar 10 05:38:08 vm02 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.b.service: Consumed 3.745s CPU time. 2026-03-10T05:38:08.505 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:08 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-c[50916]: 2026-03-10T05:38:08.295+0000 7fb8a5c0a640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.c -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T05:38:08.505 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:08 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-c[50916]: 2026-03-10T05:38:08.295+0000 7fb8a5c0a640 -1 mon.c@1(peon) e3 *** Got Signal Terminated *** 2026-03-10T05:38:08.505 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:08 vm05 podman[76696]: 2026-03-10 05:38:08.349999164 +0000 UTC m=+0.129049485 container died 88c3ee012d9d6fb71907cc0e2a0471d3d9251e7022cb2e763986436ea0af6b9c (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-c, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-10T05:38:08.505 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:08 vm05 podman[76696]: 2026-03-10 05:38:08.402063814 +0000 UTC m=+0.181114135 container remove 88c3ee012d9d6fb71907cc0e2a0471d3d9251e7022cb2e763986436ea0af6b9c (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-c, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, OSD_FLAVOR=default, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-10T05:38:08.505 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:08 vm05 bash[76696]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-c 2026-03-10T05:38:08.505 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 10 05:38:08 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-5[54445]: 2026-03-10T05:38:08.278+0000 7fc7d2f6f640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.5 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T05:38:08.505 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 10 05:38:08 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-5[54445]: 2026-03-10T05:38:08.278+0000 7fc7d2f6f640 -1 osd.5 58 *** Got signal Terminated *** 2026-03-10T05:38:08.505 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 10 05:38:08 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-5[54445]: 2026-03-10T05:38:08.278+0000 7fc7d2f6f640 -1 osd.5 58 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T05:38:08.505 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:38:08 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-6[59586]: 2026-03-10T05:38:08.275+0000 7f3ecbebf640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.6 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T05:38:08.505 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:38:08 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-6[59586]: 2026-03-10T05:38:08.275+0000 7f3ecbebf640 -1 osd.6 58 *** Got signal Terminated *** 2026-03-10T05:38:08.505 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:38:08 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-6[59586]: 2026-03-10T05:38:08.275+0000 7f3ecbebf640 -1 osd.6 58 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T05:38:08.505 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:38:08 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-7[64413]: 2026-03-10T05:38:08.250+0000 7f830c8d7640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.7 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T05:38:08.505 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:38:08 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-7[64413]: 2026-03-10T05:38:08.250+0000 7f830c8d7640 -1 osd.7 58 *** Got signal Terminated *** 2026-03-10T05:38:08.505 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:38:08 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-7[64413]: 2026-03-10T05:38:08.250+0000 7f830c8d7640 -1 osd.7 58 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T05:38:08.506 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:08 vm05 systemd[1]: Stopping Ceph iscsi.iscsi.b for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:38:08.506 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:08 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug Shutdown received 2026-03-10T05:38:08.506 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:08 vm05 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b[68246]: debug No gateway configuration to remove on this host (vm05.local) 2026-03-10T05:38:08.506 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:08 vm05 podman[76751]: 2026-03-10 05:38:08.478654208 +0000 UTC m=+0.108467218 container died 3d96d5824f4f6b2e91580e0a3428f84f89b601dccae9c0f90e36b2652319e26d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b, CEPH_REF=squid, io.buildah.version=1.41.3, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-10T05:38:08.588 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:38:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-0[57438]: 2026-03-10T05:38:08.450+0000 7f8b5351d640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T05:38:08.588 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:38:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-0[57438]: 2026-03-10T05:38:08.450+0000 7f8b5351d640 -1 osd.0 58 *** Got signal Terminated *** 2026-03-10T05:38:08.588 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:38:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-0[57438]: 2026-03-10T05:38:08.450+0000 7f8b5351d640 -1 osd.0 58 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T05:38:08.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:08 vm01 systemd[1]: Stopping Ceph iscsi.iscsi.a for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba... 2026-03-10T05:38:08.589 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:38:08 vm01 podman[80205]: 2026-03-10 05:38:08.472708588 +0000 UTC m=+0.289684600 container died 031e5c93146d7d6adf57c8a41daab78090cc8ed619b7bb2d9c824ff023deed7f (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, org.label-schema.schema-version=1.0) 2026-03-10T05:38:08.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:38:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-a[47937]: 2026-03-10T05:38:08.533+0000 7f3c1bd11640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T05:38:08.589 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:38:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-a[47937]: 2026-03-10T05:38:08.533+0000 7f3c1bd11640 -1 mon.a@0(leader) e3 *** Got Signal Terminated *** 2026-03-10T05:38:08.589 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:38:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-1[62386]: 2026-03-10T05:38:08.382+0000 7f3903e8d640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T05:38:08.589 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:38:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-1[62386]: 2026-03-10T05:38:08.382+0000 7f3903e8d640 -1 osd.1 58 *** Got signal Terminated *** 2026-03-10T05:38:08.589 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:38:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-1[62386]: 2026-03-10T05:38:08.382+0000 7f3903e8d640 -1 osd.1 58 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T05:38:08.589 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:38:08 vm01 podman[80227]: 2026-03-10 05:38:08.535001398 +0000 UTC m=+0.251984160 container died f279e724116328151575dd18ef990dceac1c1d728ef400279b88b3b3d8fbbc57 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-1, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , ceph=True, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2) 2026-03-10T05:38:08.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:08 vm05 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.c.service: Deactivated successfully. 2026-03-10T05:38:08.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:08 vm05 systemd[1]: Stopped Ceph mon.c for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:38:08.806 INFO:journalctl@ceph.mon.c.vm05.stdout:Mar 10 05:38:08 vm05 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.c.service: Consumed 3.945s CPU time. 2026-03-10T05:38:08.806 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:08 vm05 podman[76751]: 2026-03-10 05:38:08.53858576 +0000 UTC m=+0.168398770 container remove 3d96d5824f4f6b2e91580e0a3428f84f89b601dccae9c0f90e36b2652319e26d (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b, ceph=True, org.label-schema.schema-version=1.0, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3) 2026-03-10T05:38:08.806 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:08 vm05 bash[76751]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-b 2026-03-10T05:38:08.806 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:08 vm05 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@iscsi.iscsi.b.service: Deactivated successfully. 2026-03-10T05:38:08.806 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:08 vm05 systemd[1]: Stopped Ceph iscsi.iscsi.b for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:38:08.806 INFO:journalctl@ceph.iscsi.iscsi.b.vm05.stdout:Mar 10 05:38:08 vm05 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@iscsi.iscsi.b.service: Consumed 4.936s CPU time. 2026-03-10T05:38:08.846 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:38:08 vm01 podman[80255]: 2026-03-10 05:38:08.589326496 +0000 UTC m=+0.222679721 container died 55aa508bdfbd90721d88fc5e7e914984ca0e68320df3ea65a18087cb1b556813 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-0, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, org.label-schema.build-date=20260223, ceph=True, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3) 2026-03-10T05:38:08.847 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:38:08 vm01 podman[80255]: 2026-03-10 05:38:08.680895552 +0000 UTC m=+0.314248778 container remove 55aa508bdfbd90721d88fc5e7e914984ca0e68320df3ea65a18087cb1b556813 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-0, io.buildah.version=1.41.3, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2, OSD_FLAVOR=default, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-10T05:38:08.847 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:38:08 vm01 bash[80255]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-0 2026-03-10T05:38:08.847 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:08 vm01 podman[80261]: 2026-03-10 05:38:08.63701511 +0000 UTC m=+0.210594122 container died 11f18971109c816585547659c764de6b76b565ab362f65c419b1640fa99ce169 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-a, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-10T05:38:08.847 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug Shutdown received 2026-03-10T05:38:08.847 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:38:08 vm01 podman[80205]: 2026-03-10 05:38:08.649906778 +0000 UTC m=+0.466882790 container remove 031e5c93146d7d6adf57c8a41daab78090cc8ed619b7bb2d9c824ff023deed7f (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x, ceph=True, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T05:38:08.847 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:38:08 vm01 bash[80205]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mgr-x 2026-03-10T05:38:08.847 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:38:08 vm01 podman[80277]: 2026-03-10 05:38:08.690364743 +0000 UTC m=+0.222433129 container remove 11f18971109c816585547659c764de6b76b565ab362f65c419b1640fa99ce169 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-a, ceph=True, OSD_FLAVOR=default, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T05:38:08.847 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:38:08 vm01 bash[80277]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-mon-a 2026-03-10T05:38:08.847 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:38:08 vm01 podman[80227]: 2026-03-10 05:38:08.635601692 +0000 UTC m=+0.352584465 container remove f279e724116328151575dd18ef990dceac1c1d728ef400279b88b3b3d8fbbc57 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-1, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, ceph=True) 2026-03-10T05:38:08.847 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:38:08 vm01 bash[80227]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-1 2026-03-10T05:38:09.127 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:08 vm01 ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a[66495]: debug No gateway configuration to remove on this host (vm01.local) 2026-03-10T05:38:09.128 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:08 vm01 podman[80367]: 2026-03-10 05:38:08.964254999 +0000 UTC m=+0.204951486 container died 87eedcfb11be985407e933412b07619f5f228f5d72e96240fa07cd7a9315a318 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, CEPH_REF=squid, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-10T05:38:09.128 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:08 vm01 podman[80367]: 2026-03-10 05:38:08.988742179 +0000 UTC m=+0.229438676 container remove 87eedcfb11be985407e933412b07619f5f228f5d72e96240fa07cd7a9315a318 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, CEPH_REF=squid, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T05:38:09.128 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:08 vm01 bash[80367]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-iscsi-iscsi-a 2026-03-10T05:38:09.128 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:38:08 vm01 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mgr.x.service: Deactivated successfully. 2026-03-10T05:38:09.128 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:38:08 vm01 systemd[1]: Stopped Ceph mgr.x for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:38:09.128 INFO:journalctl@ceph.mgr.x.vm01.stdout:Mar 10 05:38:08 vm01 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mgr.x.service: Consumed 18.761s CPU time. 2026-03-10T05:38:09.128 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:38:08 vm01 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.a.service: Deactivated successfully. 2026-03-10T05:38:09.128 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:38:08 vm01 systemd[1]: Stopped Ceph mon.a for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:38:09.128 INFO:journalctl@ceph.mon.a.vm01.stdout:Mar 10 05:38:08 vm01 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.a.service: Consumed 5.341s CPU time. 2026-03-10T05:38:09.588 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:38:09 vm01 bash[80492]: Error: statfs /var/log/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba: no such file or directory 2026-03-10T05:38:09.588 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:38:09 vm01 podman[80492]: 2026-03-10 05:38:09.139992202 +0000 UTC m=+0.021969457 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T05:38:09.588 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:38:09 vm01 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.0.service: Deactivated successfully. 2026-03-10T05:38:09.588 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:38:09 vm01 systemd[1]: Stopped Ceph osd.0 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:38:09.588 INFO:journalctl@ceph.osd.0.vm01.stdout:Mar 10 05:38:09 vm01 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.0.service: Consumed 2.108s CPU time. 2026-03-10T05:38:09.588 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:38:09 vm01 bash[80509]: Error: statfs /var/log/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba: no such file or directory 2026-03-10T05:38:09.588 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:38:09 vm01 podman[80509]: 2026-03-10 05:38:09.182243334 +0000 UTC m=+0.014730723 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T05:38:09.588 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:38:09 vm01 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.1.service: Deactivated successfully. 2026-03-10T05:38:09.588 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:38:09 vm01 systemd[1]: Stopped Ceph osd.1 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:38:09.588 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 10 05:38:09 vm01 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.1.service: Consumed 2.149s CPU time. 2026-03-10T05:38:09.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:09 vm01 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@iscsi.iscsi.a.service: Deactivated successfully. 2026-03-10T05:38:09.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:09 vm01 systemd[1]: Stopped Ceph iscsi.iscsi.a for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:38:09.588 INFO:journalctl@ceph.iscsi.iscsi.a.vm01.stdout:Mar 10 05:38:09 vm01 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@iscsi.iscsi.a.service: Consumed 12.402s CPU time. 2026-03-10T05:38:13.153 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:38:13 vm02 podman[73622]: 2026-03-10 05:38:13.150450476 +0000 UTC m=+5.056321731 container died c69c0cfb0df9a1ed21e1e9bd92fdfe370d643d26a5be321e4f38e6690114d350 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-3, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, org.label-schema.build-date=20260223, CEPH_REF=squid, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-10T05:38:13.556 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:38:13 vm05 podman[76695]: 2026-03-10 05:38:13.434927424 +0000 UTC m=+5.232904424 container died 240f37621334619d2ccf693b9150013fef6ec5fa25aaf2584992815d6563df36 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-6, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-10T05:38:13.556 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:38:13 vm05 podman[76695]: 2026-03-10 05:38:13.45859341 +0000 UTC m=+5.256570410 container remove 240f37621334619d2ccf693b9150013fef6ec5fa25aaf2584992815d6563df36 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-6, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-10T05:38:13.556 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:38:13 vm05 bash[76695]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-6 2026-03-10T05:38:13.556 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:38:13 vm05 podman[76669]: 2026-03-10 05:38:13.293156344 +0000 UTC m=+5.126329246 container died cf2813e55166cdde0d949867ab66d55bccf3dedb4ab1805464fd27e38b4f4ce9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-7, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-10T05:38:13.556 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:38:13 vm05 podman[76669]: 2026-03-10 05:38:13.425420483 +0000 UTC m=+5.258593385 container remove cf2813e55166cdde0d949867ab66d55bccf3dedb4ab1805464fd27e38b4f4ce9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-7, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-10T05:38:13.556 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:38:13 vm05 bash[76669]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-7 2026-03-10T05:38:13.556 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 10 05:38:13 vm05 podman[76694]: 2026-03-10 05:38:13.336903359 +0000 UTC m=+5.130939132 container died 0f019ca23e77a54cc651c32d915a55fb51ef7488cb3f0d7c399c8d409c4e2d39 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-5, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, ceph=True, io.buildah.version=1.41.3) 2026-03-10T05:38:13.556 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 10 05:38:13 vm05 podman[76694]: 2026-03-10 05:38:13.393340371 +0000 UTC m=+5.187376144 container remove 0f019ca23e77a54cc651c32d915a55fb51ef7488cb3f0d7c399c8d409c4e2d39 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-5, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS) 2026-03-10T05:38:13.556 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 10 05:38:13 vm05 bash[76694]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-5 2026-03-10T05:38:13.620 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:38:13 vm02 podman[73602]: 2026-03-10 05:38:13.193071819 +0000 UTC m=+5.124179494 container died cd32e7233680b6125a3e28b8d3f1132d62d3c9dd7f9af005086a8a3674806581 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-2, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, ceph=True) 2026-03-10T05:38:13.620 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:38:13 vm02 podman[73602]: 2026-03-10 05:38:13.227163947 +0000 UTC m=+5.158271622 container remove cd32e7233680b6125a3e28b8d3f1132d62d3c9dd7f9af005086a8a3674806581 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-2, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-10T05:38:13.620 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:38:13 vm02 bash[73602]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-2 2026-03-10T05:38:13.620 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:38:13 vm02 podman[73622]: 2026-03-10 05:38:13.222618287 +0000 UTC m=+5.128489542 container remove c69c0cfb0df9a1ed21e1e9bd92fdfe370d643d26a5be321e4f38e6690114d350 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223) 2026-03-10T05:38:13.620 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:38:13 vm02 bash[73622]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-3 2026-03-10T05:38:13.620 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:38:13 vm02 bash[73840]: Error: statfs /var/log/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba: no such file or directory 2026-03-10T05:38:13.620 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:38:13 vm02 podman[73641]: 2026-03-10 05:38:13.279346536 +0000 UTC m=+5.094142428 container died dfc2f16e1637acdadfdfdb78dec3b8dd37c3ae89259504618326bf9f622d69b5 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-4, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, ceph=True, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-10T05:38:13.620 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:38:13 vm02 podman[73641]: 2026-03-10 05:38:13.303436651 +0000 UTC m=+5.118232543 container remove dfc2f16e1637acdadfdfdb78dec3b8dd37c3ae89259504618326bf9f622d69b5 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-4, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, ceph=True, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-10T05:38:13.620 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:38:13 vm02 bash[73641]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba-osd-4 2026-03-10T05:38:13.898 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:38:13 vm02 bash[73874]: Error: statfs /var/log/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba: no such file or directory 2026-03-10T05:38:13.898 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:38:13 vm02 podman[73874]: 2026-03-10 05:38:13.69179067 +0000 UTC m=+0.010099366 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T05:38:13.898 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:38:13 vm02 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.2.service: Deactivated successfully. 2026-03-10T05:38:13.899 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:38:13 vm02 systemd[1]: Stopped Ceph osd.2 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:38:13.899 INFO:journalctl@ceph.osd.2.vm02.stdout:Mar 10 05:38:13 vm02 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.2.service: Consumed 2.373s CPU time. 2026-03-10T05:38:13.899 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:38:13 vm02 podman[73840]: 2026-03-10 05:38:13.611515283 +0000 UTC m=+0.028052965 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T05:38:13.899 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:38:13 vm02 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.3.service: Deactivated successfully. 2026-03-10T05:38:13.899 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:38:13 vm02 systemd[1]: Stopped Ceph osd.3 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:38:13.899 INFO:journalctl@ceph.osd.3.vm02.stdout:Mar 10 05:38:13 vm02 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.3.service: Consumed 2.736s CPU time. 2026-03-10T05:38:13.899 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:38:13 vm02 bash[73858]: Error: statfs /var/log/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba: no such file or directory 2026-03-10T05:38:13.899 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:38:13 vm02 podman[73858]: 2026-03-10 05:38:13.654752378 +0000 UTC m=+0.016919011 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T05:38:13.899 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:38:13 vm02 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.4.service: Deactivated successfully. 2026-03-10T05:38:13.899 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:38:13 vm02 systemd[1]: Stopped Ceph osd.4 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:38:13.899 INFO:journalctl@ceph.osd.4.vm02.stdout:Mar 10 05:38:13 vm02 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.4.service: Consumed 1.620s CPU time. 2026-03-10T05:38:14.055 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 10 05:38:13 vm05 bash[77050]: Error: statfs /var/log/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba: no such file or directory 2026-03-10T05:38:14.056 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 10 05:38:13 vm05 podman[77050]: 2026-03-10 05:38:13.813820916 +0000 UTC m=+0.035912916 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T05:38:14.056 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 10 05:38:13 vm05 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.5.service: Deactivated successfully. 2026-03-10T05:38:14.056 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 10 05:38:13 vm05 systemd[1]: Stopped Ceph osd.5 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:38:14.056 INFO:journalctl@ceph.osd.5.vm05.stdout:Mar 10 05:38:13 vm05 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.5.service: Consumed 2.022s CPU time. 2026-03-10T05:38:14.056 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:38:13 vm05 bash[77084]: Error: statfs /var/log/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba: no such file or directory 2026-03-10T05:38:14.056 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:38:13 vm05 podman[77084]: 2026-03-10 05:38:13.835147716 +0000 UTC m=+0.017502663 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T05:38:14.056 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:38:13 vm05 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.6.service: Deactivated successfully. 2026-03-10T05:38:14.056 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:38:13 vm05 systemd[1]: Stopped Ceph osd.6 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:38:14.056 INFO:journalctl@ceph.osd.6.vm05.stdout:Mar 10 05:38:13 vm05 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.6.service: Consumed 1.716s CPU time. 2026-03-10T05:38:14.056 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:38:13 vm05 bash[77015]: Error: statfs /var/log/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba: no such file or directory 2026-03-10T05:38:14.056 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:38:13 vm05 podman[77015]: 2026-03-10 05:38:13.77053295 +0000 UTC m=+0.023201629 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T05:38:14.056 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:38:13 vm05 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.7.service: Deactivated successfully. 2026-03-10T05:38:14.056 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:38:13 vm05 systemd[1]: Stopped Ceph osd.7 for dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba. 2026-03-10T05:38:14.056 INFO:journalctl@ceph.osd.7.vm05.stdout:Mar 10 05:38:13 vm05 systemd[1]: ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.7.service: Consumed 1.509s CPU time. 2026-03-10T05:38:14.165 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-10T05:38:14.166 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /sys 2026-03-10T05:38:14.166 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /proc 2026-03-10T05:38:14.166 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /mnt 2026-03-10T05:38:14.166 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /var/tmp 2026-03-10T05:38:14.166 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /home 2026-03-10T05:38:14.166 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /root 2026-03-10T05:38:14.166 INFO:teuthology.orchestra.run.vm05.stdout:skipping the directory /tmp 2026-03-10T05:38:14.166 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:14.175 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-10T05:38:14.177 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-10T05:38:14.177 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /sys 2026-03-10T05:38:14.177 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /proc 2026-03-10T05:38:14.177 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /mnt 2026-03-10T05:38:14.177 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /var/tmp 2026-03-10T05:38:14.177 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /home 2026-03-10T05:38:14.177 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /root 2026-03-10T05:38:14.178 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /tmp 2026-03-10T05:38:14.178 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:14.186 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-10T05:38:14.194 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T05:38:14.195 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T05:38:14.201 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T05:38:14.201 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T05:38:14.204 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T05:38:14.207 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-10T05:38:14.207 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-10T05:38:14.207 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /sys 2026-03-10T05:38:14.207 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /proc 2026-03-10T05:38:14.207 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /mnt 2026-03-10T05:38:14.207 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /var/tmp 2026-03-10T05:38:14.207 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /home 2026-03-10T05:38:14.207 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /root 2026-03-10T05:38:14.207 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /tmp 2026-03-10T05:38:14.207 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:14.209 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T05:38:14.209 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-10T05:38:14.212 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-10T05:38:14.212 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-10T05:38:14.215 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-10T05:38:14.215 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-10T05:38:14.215 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-10T05:38:14.216 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-10T05:38:14.217 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-10T05:38:14.219 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-10T05:38:14.220 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-10T05:38:14.229 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-10T05:38:14.232 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-10T05:38:14.232 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-10T05:38:14.234 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-10T05:38:14.234 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-10T05:38:14.235 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T05:38:14.236 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T05:38:14.236 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-10T05:38:14.238 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-10T05:38:14.239 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-10T05:38:14.241 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-10T05:38:14.242 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-10T05:38:14.245 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T05:38:14.247 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-10T05:38:14.247 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-10T05:38:14.248 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-10T05:38:14.251 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-10T05:38:14.254 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-10T05:38:14.255 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-10T05:38:14.255 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-10T05:38:14.256 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-10T05:38:14.256 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-10T05:38:14.260 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-10T05:38:14.260 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T05:38:14.261 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-10T05:38:14.261 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T05:38:14.272 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-10T05:38:14.275 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-10T05:38:14.278 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-10T05:38:14.280 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-10T05:38:14.283 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-10T05:38:14.288 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-10T05:38:14.298 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-10T05:38:14.303 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-10T05:38:14.303 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T05:38:14.358 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T05:38:14.358 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-10T05:38:14.358 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T05:38:14.358 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-10T05:38:14.358 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-10T05:38:14.358 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-10T05:38:14.358 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-10T05:38:14.358 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T05:38:14.358 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-10T05:38:14.358 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-10T05:38:14.359 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-10T05:38:14.360 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-10T05:38:14.361 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T05:38:14.361 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-10T05:38:14.361 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T05:38:14.361 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-10T05:38:14.361 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-10T05:38:14.361 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-10T05:38:14.361 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-10T05:38:14.361 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T05:38:14.361 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-10T05:38:14.361 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-10T05:38:14.362 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-10T05:38:14.362 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-10T05:38:14.363 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-10T05:38:14.366 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-10T05:38:14.367 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-10T05:38:14.419 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-10T05:38:14.420 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-10T05:38:14.421 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-10T05:38:14.422 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-10T05:38:14.423 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T05:38:14.450 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T05:38:14.451 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T05:38:14.452 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T05:38:14.453 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.453 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:14.453 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T05:38:14.455 INFO:teuthology.orchestra.run.vm05.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T05:38:14.456 INFO:teuthology.orchestra.run.vm05.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:14.457 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T05:38:14.513 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-10T05:38:14.514 INFO:teuthology.orchestra.run.vm01.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:14.515 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:14.655 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:14.656 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:14.656 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-10T05:38:14.656 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:14.656 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-10T05:38:14.656 INFO:teuthology.orchestra.run.vm02.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-10T05:38:14.656 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:14.656 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-10T05:38:14.656 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:14.656 INFO:teuthology.orchestra.run.vm02.stdout:Remove 1 Package 2026-03-10T05:38:14.656 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:14.656 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 775 k 2026-03-10T05:38:14.656 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-10T05:38:14.658 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-10T05:38:14.658 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-10T05:38:14.659 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-10T05:38:14.659 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-10T05:38:14.677 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:14.677 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-10T05:38:14.677 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T05:38:14.677 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:14.677 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-10T05:38:14.677 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:14.677 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-10T05:38:14.677 INFO:teuthology.orchestra.run.vm05.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-10T05:38:14.677 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:14.677 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-10T05:38:14.677 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:14.677 INFO:teuthology.orchestra.run.vm05.stdout:Remove 1 Package 2026-03-10T05:38:14.678 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:14.678 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 775 k 2026-03-10T05:38:14.678 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-10T05:38:14.679 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-10T05:38:14.680 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-10T05:38:14.681 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-10T05:38:14.681 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-10T05:38:14.699 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-10T05:38:14.699 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T05:38:14.733 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:14.734 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:14.734 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-10T05:38:14.734 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:14.734 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-10T05:38:14.734 INFO:teuthology.orchestra.run.vm01.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-10T05:38:14.734 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:14.734 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-10T05:38:14.734 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:14.734 INFO:teuthology.orchestra.run.vm01.stdout:Remove 1 Package 2026-03-10T05:38:14.734 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:14.734 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 775 k 2026-03-10T05:38:14.735 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-10T05:38:14.736 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-10T05:38:14.736 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-10T05:38:14.738 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-10T05:38:14.738 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-10T05:38:14.756 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-10T05:38:14.756 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T05:38:14.801 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T05:38:14.816 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T05:38:14.847 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T05:38:14.847 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:14.847 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-10T05:38:14.848 INFO:teuthology.orchestra.run.vm02.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.848 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:14.848 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:14.861 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T05:38:14.861 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:14.861 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-10T05:38:14.861 INFO:teuthology.orchestra.run.vm05.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.861 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:14.861 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:14.870 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T05:38:14.927 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T05:38:14.927 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:14.927 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-10T05:38:14.927 INFO:teuthology.orchestra.run.vm01.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T05:38:14.927 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:14.927 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:15.028 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-immutable-object-cache 2026-03-10T05:38:15.028 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:15.030 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:15.031 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:15.031 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:15.055 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: ceph-immutable-object-cache 2026-03-10T05:38:15.055 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:15.058 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:15.059 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:15.059 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:15.128 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-immutable-object-cache 2026-03-10T05:38:15.129 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:15.131 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:15.132 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:15.132 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:15.224 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr 2026-03-10T05:38:15.224 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:15.224 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:15.224 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:15.224 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:15.242 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: ceph-mgr 2026-03-10T05:38:15.242 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:15.245 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:15.245 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:15.245 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:15.317 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr 2026-03-10T05:38:15.318 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:15.320 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:15.320 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:15.320 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:15.418 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr-dashboard 2026-03-10T05:38:15.418 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:15.420 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:15.421 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:15.421 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:15.423 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: ceph-mgr-dashboard 2026-03-10T05:38:15.423 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:15.425 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:15.426 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:15.426 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:15.499 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-dashboard 2026-03-10T05:38:15.499 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:15.499 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:15.499 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:15.499 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:15.590 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-10T05:38:15.591 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:15.592 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:15.593 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-10T05:38:15.593 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:15.593 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:15.593 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:15.595 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:15.596 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:15.596 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:15.676 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-10T05:38:15.676 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:15.678 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:15.679 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:15.679 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:15.765 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: ceph-mgr-rook 2026-03-10T05:38:15.765 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:15.766 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr-rook 2026-03-10T05:38:15.766 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:15.767 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:15.767 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:15.767 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:15.769 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:15.769 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:15.769 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:15.850 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-rook 2026-03-10T05:38:15.850 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:15.852 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:15.853 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:15.853 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:15.936 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: ceph-mgr-cephadm 2026-03-10T05:38:15.936 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:15.938 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:15.939 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:15.939 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:15.943 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr-cephadm 2026-03-10T05:38:15.943 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:15.945 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:15.946 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:15.946 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:16.024 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-cephadm 2026-03-10T05:38:16.024 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:16.026 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:16.027 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:16.027 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:16.118 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:16.118 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:16.118 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-10T05:38:16.118 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:16.118 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-10T05:38:16.118 INFO:teuthology.orchestra.run.vm05.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-10T05:38:16.118 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:16.118 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-10T05:38:16.118 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:16.118 INFO:teuthology.orchestra.run.vm05.stdout:Remove 1 Package 2026-03-10T05:38:16.118 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:16.118 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 3.6 M 2026-03-10T05:38:16.118 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-10T05:38:16.120 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-10T05:38:16.120 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-10T05:38:16.130 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-10T05:38:16.130 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-10T05:38:16.143 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:16.143 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:16.143 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-10T05:38:16.143 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:16.143 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-10T05:38:16.143 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-10T05:38:16.143 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:16.143 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-10T05:38:16.143 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:16.143 INFO:teuthology.orchestra.run.vm02.stdout:Remove 1 Package 2026-03-10T05:38:16.143 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:16.143 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 3.6 M 2026-03-10T05:38:16.143 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-10T05:38:16.145 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-10T05:38:16.145 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-10T05:38:16.155 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-10T05:38:16.155 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-10T05:38:16.159 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-10T05:38:16.174 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T05:38:16.184 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-10T05:38:16.199 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T05:38:16.211 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:16.211 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:16.211 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-10T05:38:16.211 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:16.211 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-10T05:38:16.211 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-10T05:38:16.211 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:16.211 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-10T05:38:16.211 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:16.211 INFO:teuthology.orchestra.run.vm01.stdout:Remove 1 Package 2026-03-10T05:38:16.211 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:16.211 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 3.6 M 2026-03-10T05:38:16.211 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-10T05:38:16.213 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-10T05:38:16.213 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-10T05:38:16.223 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-10T05:38:16.223 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-10T05:38:16.245 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T05:38:16.249 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-10T05:38:16.264 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T05:38:16.269 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T05:38:16.290 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T05:38:16.290 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:16.290 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-10T05:38:16.290 INFO:teuthology.orchestra.run.vm05.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:16.290 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:16.290 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:16.322 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T05:38:16.322 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:16.322 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-10T05:38:16.322 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:16.322 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:16.322 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:16.342 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T05:38:16.382 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T05:38:16.382 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:16.382 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-10T05:38:16.382 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:16.382 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:16.382 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:16.487 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: ceph-volume 2026-03-10T05:38:16.487 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:16.489 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:16.490 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:16.490 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:16.511 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-volume 2026-03-10T05:38:16.512 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:16.513 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:16.514 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:16.514 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:16.563 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-volume 2026-03-10T05:38:16.563 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:16.565 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:16.566 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:16.566 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:16.674 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:16.675 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:16.675 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repo Size 2026-03-10T05:38:16.675 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:16.675 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-10T05:38:16.675 INFO:teuthology.orchestra.run.vm05.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-10T05:38:16.675 INFO:teuthology.orchestra.run.vm05.stdout:Removing dependent packages: 2026-03-10T05:38:16.675 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-10T05:38:16.675 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:16.675 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-10T05:38:16.675 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:16.675 INFO:teuthology.orchestra.run.vm05.stdout:Remove 2 Packages 2026-03-10T05:38:16.675 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:16.675 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 610 k 2026-03-10T05:38:16.675 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-10T05:38:16.677 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-10T05:38:16.677 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-10T05:38:16.688 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-10T05:38:16.688 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-10T05:38:16.705 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:16.705 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:16.705 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repo Size 2026-03-10T05:38:16.705 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:16.705 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-10T05:38:16.705 INFO:teuthology.orchestra.run.vm02.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-10T05:38:16.705 INFO:teuthology.orchestra.run.vm02.stdout:Removing dependent packages: 2026-03-10T05:38:16.705 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-10T05:38:16.705 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:16.706 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-10T05:38:16.706 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:16.706 INFO:teuthology.orchestra.run.vm02.stdout:Remove 2 Packages 2026-03-10T05:38:16.706 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:16.706 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 610 k 2026-03-10T05:38:16.706 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-10T05:38:16.707 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-10T05:38:16.707 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-10T05:38:16.714 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-10T05:38:16.716 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:16.719 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-10T05:38:16.719 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-10T05:38:16.730 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T05:38:16.746 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-10T05:38:16.748 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:16.758 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:16.759 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:16.759 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repo Size 2026-03-10T05:38:16.759 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:16.759 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-10T05:38:16.759 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-10T05:38:16.759 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-10T05:38:16.759 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-10T05:38:16.759 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:16.759 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-10T05:38:16.759 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:16.759 INFO:teuthology.orchestra.run.vm01.stdout:Remove 2 Packages 2026-03-10T05:38:16.759 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:16.759 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 610 k 2026-03-10T05:38:16.759 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-10T05:38:16.761 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-10T05:38:16.761 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-10T05:38:16.762 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T05:38:16.771 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-10T05:38:16.771 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-10T05:38:16.796 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T05:38:16.797 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:16.798 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-10T05:38:16.801 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:16.816 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T05:38:16.836 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T05:38:16.836 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:16.846 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T05:38:16.846 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:16.846 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-10T05:38:16.846 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:16.846 INFO:teuthology.orchestra.run.vm05.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:16.846 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:16.846 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:16.889 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T05:38:16.889 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:16.889 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-10T05:38:16.889 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:16.889 INFO:teuthology.orchestra.run.vm02.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:16.889 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:16.889 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:16.889 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T05:38:16.889 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T05:38:16.937 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T05:38:16.937 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:16.937 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-10T05:38:16.937 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:16.937 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:16.937 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:16.937 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:17.052 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repo Size 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout:Removing dependent packages: 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout:Removing unused dependencies: 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout:Remove 3 Packages 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 3.7 M 2026-03-10T05:38:17.053 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-10T05:38:17.055 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-10T05:38:17.055 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-10T05:38:17.073 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-10T05:38:17.073 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-10T05:38:17.074 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repo Size 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout:Removing dependent packages: 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout:Remove 3 Packages 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 3.7 M 2026-03-10T05:38:17.075 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-10T05:38:17.077 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-10T05:38:17.077 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-10T05:38:17.094 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-10T05:38:17.094 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-10T05:38:17.106 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-10T05:38:17.109 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T05:38:17.110 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T05:38:17.110 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T05:38:17.124 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-10T05:38:17.126 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T05:38:17.127 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T05:38:17.128 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T05:38:17.142 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repo Size 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout:Remove 3 Packages 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:17.143 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 3.7 M 2026-03-10T05:38:17.144 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-10T05:38:17.145 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-10T05:38:17.145 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-10T05:38:17.161 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-10T05:38:17.162 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-10T05:38:17.173 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T05:38:17.173 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T05:38:17.173 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T05:38:17.193 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T05:38:17.193 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T05:38:17.193 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T05:38:17.194 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-10T05:38:17.196 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T05:38:17.198 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T05:38:17.198 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T05:38:17.209 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T05:38:17.209 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:17.209 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-10T05:38:17.209 INFO:teuthology.orchestra.run.vm05.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.209 INFO:teuthology.orchestra.run.vm05.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.209 INFO:teuthology.orchestra.run.vm05.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.209 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:17.210 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:17.232 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T05:38:17.232 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:17.232 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-10T05:38:17.232 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.232 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.232 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.232 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:17.232 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:17.264 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T05:38:17.265 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T05:38:17.265 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T05:38:17.302 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T05:38:17.302 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:17.302 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-10T05:38:17.302 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.302 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.302 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.302 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:17.302 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:17.389 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: libcephfs-devel 2026-03-10T05:38:17.389 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:17.391 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:17.392 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:17.392 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:17.407 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: libcephfs-devel 2026-03-10T05:38:17.407 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:17.409 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:17.410 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:17.410 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:17.465 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: libcephfs-devel 2026-03-10T05:38:17.465 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:17.468 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:17.469 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:17.469 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:17.579 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: Package Arch Version Repository Size 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout:Removing: 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout:Removing dependent packages: 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout:Removing unused dependencies: 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout:Transaction Summary 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout:================================================================================ 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout:Remove 20 Packages 2026-03-10T05:38:17.581 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:17.582 INFO:teuthology.orchestra.run.vm05.stdout:Freed space: 79 M 2026-03-10T05:38:17.582 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction check 2026-03-10T05:38:17.585 INFO:teuthology.orchestra.run.vm05.stdout:Transaction check succeeded. 2026-03-10T05:38:17.586 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction test 2026-03-10T05:38:17.597 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout:Removing dependent packages: 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-10T05:38:17.598 INFO:teuthology.orchestra.run.vm02.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-10T05:38:17.599 INFO:teuthology.orchestra.run.vm02.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T05:38:17.599 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-10T05:38:17.599 INFO:teuthology.orchestra.run.vm02.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-10T05:38:17.599 INFO:teuthology.orchestra.run.vm02.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-10T05:38:17.599 INFO:teuthology.orchestra.run.vm02.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-10T05:38:17.599 INFO:teuthology.orchestra.run.vm02.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-10T05:38:17.599 INFO:teuthology.orchestra.run.vm02.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-10T05:38:17.599 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:17.599 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-10T05:38:17.599 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-10T05:38:17.599 INFO:teuthology.orchestra.run.vm02.stdout:Remove 20 Packages 2026-03-10T05:38:17.599 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:17.599 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 79 M 2026-03-10T05:38:17.599 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-10T05:38:17.603 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-10T05:38:17.603 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-10T05:38:17.609 INFO:teuthology.orchestra.run.vm05.stdout:Transaction test succeeded. 2026-03-10T05:38:17.609 INFO:teuthology.orchestra.run.vm05.stdout:Running transaction 2026-03-10T05:38:17.627 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-10T05:38:17.627 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-10T05:38:17.650 INFO:teuthology.orchestra.run.vm05.stdout: Preparing : 1/1 2026-03-10T05:38:17.652 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-10T05:38:17.654 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-10T05:38:17.658 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-10T05:38:17.658 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T05:38:17.664 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:17.665 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:17.665 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-10T05:38:17.665 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout:Remove 20 Packages 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 79 M 2026-03-10T05:38:17.666 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-10T05:38:17.670 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-10T05:38:17.670 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-10T05:38:17.671 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T05:38:17.672 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-10T05:38:17.674 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-10T05:38:17.675 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-10T05:38:17.676 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-10T05:38:17.677 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-10T05:38:17.678 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T05:38:17.679 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-10T05:38:17.680 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-10T05:38:17.680 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T05:38:17.682 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-10T05:38:17.682 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T05:38:17.694 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-10T05:38:17.694 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T05:38:17.694 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-10T05:38:17.696 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T05:38:17.696 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T05:38:17.696 INFO:teuthology.orchestra.run.vm05.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-10T05:38:17.696 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:17.696 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-10T05:38:17.698 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-10T05:38:17.700 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T05:38:17.702 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-10T05:38:17.705 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-10T05:38:17.705 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T05:38:17.710 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T05:38:17.712 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-10T05:38:17.716 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-10T05:38:17.720 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-10T05:38:17.721 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T05:38:17.721 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T05:38:17.721 INFO:teuthology.orchestra.run.vm02.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-10T05:38:17.721 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:17.722 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-10T05:38:17.725 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-10T05:38:17.727 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-10T05:38:17.728 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-10T05:38:17.730 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-10T05:38:17.735 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T05:38:17.735 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-10T05:38:17.737 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-10T05:38:17.738 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-10T05:38:17.740 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-10T05:38:17.741 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-10T05:38:17.744 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-10T05:38:17.744 INFO:teuthology.orchestra.run.vm05.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T05:38:17.745 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-10T05:38:17.745 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T05:38:17.747 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-10T05:38:17.749 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-10T05:38:17.751 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-10T05:38:17.753 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-10T05:38:17.754 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-10T05:38:17.760 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T05:38:17.762 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-10T05:38:17.765 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-10T05:38:17.767 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T05:38:17.768 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T05:38:17.769 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-10T05:38:17.771 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-10T05:38:17.771 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T05:38:17.786 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T05:38:17.786 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T05:38:17.786 INFO:teuthology.orchestra.run.vm01.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-10T05:38:17.786 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:17.801 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T05:38:17.803 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-10T05:38:17.805 INFO:teuthology.orchestra.run.vm05.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T05:38:17.805 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-10T05:38:17.805 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-10T05:38:17.805 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-10T05:38:17.805 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-10T05:38:17.805 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-10T05:38:17.805 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-10T05:38:17.806 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T05:38:17.806 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-10T05:38:17.806 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-10T05:38:17.806 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T05:38:17.806 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-10T05:38:17.806 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-10T05:38:17.806 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-10T05:38:17.806 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-10T05:38:17.806 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-10T05:38:17.806 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-10T05:38:17.806 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-10T05:38:17.806 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-10T05:38:17.806 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-10T05:38:17.807 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-10T05:38:17.810 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-10T05:38:17.813 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-10T05:38:17.816 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-10T05:38:17.818 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-10T05:38:17.819 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-10T05:38:17.821 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-10T05:38:17.836 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T05:38:17.840 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T05:38:17.840 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-10T05:38:17.841 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout:Removed: 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout: 2026-03-10T05:38:17.859 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:17.884 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-10T05:38:17.884 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:17.884 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-10T05:38:17.884 INFO:teuthology.orchestra.run.vm02.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T05:38:17.884 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-10T05:38:17.885 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:17.904 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-10T05:38:17.905 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-10T05:38:17.951 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-10T05:38:17.951 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:17.951 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-10T05:38:17.951 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T05:38:17.951 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T05:38:17.951 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T05:38:17.951 INFO:teuthology.orchestra.run.vm01.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T05:38:17.951 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T05:38:17.951 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T05:38:17.951 INFO:teuthology.orchestra.run.vm01.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.951 INFO:teuthology.orchestra.run.vm01.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.951 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T05:38:17.951 INFO:teuthology.orchestra.run.vm01.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.952 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T05:38:17.952 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T05:38:17.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.952 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.952 INFO:teuthology.orchestra.run.vm01.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-10T05:38:17.952 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.952 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T05:38:17.952 INFO:teuthology.orchestra.run.vm01.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T05:38:17.952 INFO:teuthology.orchestra.run.vm01.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T05:38:17.952 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-10T05:38:17.952 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:18.060 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: librbd1 2026-03-10T05:38:18.060 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:18.062 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:18.063 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:18.063 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:18.084 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: librbd1 2026-03-10T05:38:18.084 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:18.086 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:18.087 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:18.087 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:18.141 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: librbd1 2026-03-10T05:38:18.142 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:18.144 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:18.144 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:18.145 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:18.271 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: python3-rados 2026-03-10T05:38:18.271 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:18.273 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:18.274 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: python3-rados 2026-03-10T05:38:18.274 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:18.274 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:18.274 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:18.277 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:18.277 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:18.277 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:18.334 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rados 2026-03-10T05:38:18.334 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:18.336 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:18.337 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:18.337 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:18.449 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: python3-rgw 2026-03-10T05:38:18.449 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:18.451 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:18.452 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:18.452 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:18.453 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: python3-rgw 2026-03-10T05:38:18.453 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:18.455 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:18.456 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:18.456 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:18.508 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rgw 2026-03-10T05:38:18.509 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:18.510 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:18.511 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:18.511 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:18.621 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: python3-cephfs 2026-03-10T05:38:18.621 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:18.624 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:18.624 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:18.624 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:18.635 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: python3-cephfs 2026-03-10T05:38:18.635 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:18.638 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:18.638 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:18.638 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:18.691 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-cephfs 2026-03-10T05:38:18.691 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:18.693 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:18.693 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:18.693 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:18.815 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: python3-rbd 2026-03-10T05:38:18.816 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:18.817 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: python3-rbd 2026-03-10T05:38:18.817 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:18.818 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:18.818 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:18.819 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:18.819 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:18.820 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:18.820 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:18.869 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rbd 2026-03-10T05:38:18.869 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:18.871 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:18.871 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:18.871 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:18.984 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: rbd-fuse 2026-03-10T05:38:18.984 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:18.986 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:18.987 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:18.987 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:18.988 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: rbd-fuse 2026-03-10T05:38:18.988 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:18.991 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:18.991 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:18.991 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:19.036 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-fuse 2026-03-10T05:38:19.036 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:19.039 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:19.039 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:19.039 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:19.162 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: rbd-mirror 2026-03-10T05:38:19.162 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:19.164 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:19.165 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:19.165 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:19.168 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: rbd-mirror 2026-03-10T05:38:19.168 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:19.170 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:19.171 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:19.171 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:19.204 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-mirror 2026-03-10T05:38:19.204 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:19.206 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:19.207 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:19.207 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:19.335 INFO:teuthology.orchestra.run.vm05.stdout:No match for argument: rbd-nbd 2026-03-10T05:38:19.335 INFO:teuthology.orchestra.run.vm05.stderr:No packages marked for removal. 2026-03-10T05:38:19.337 INFO:teuthology.orchestra.run.vm05.stdout:Dependencies resolved. 2026-03-10T05:38:19.337 INFO:teuthology.orchestra.run.vm05.stdout:Nothing to do. 2026-03-10T05:38:19.337 INFO:teuthology.orchestra.run.vm05.stdout:Complete! 2026-03-10T05:38:19.338 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: rbd-nbd 2026-03-10T05:38:19.338 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-10T05:38:19.340 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-10T05:38:19.341 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-10T05:38:19.341 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-10T05:38:19.361 DEBUG:teuthology.orchestra.run.vm05:> sudo yum clean all 2026-03-10T05:38:19.362 DEBUG:teuthology.orchestra.run.vm02:> sudo yum clean all 2026-03-10T05:38:19.377 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-nbd 2026-03-10T05:38:19.377 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-10T05:38:19.379 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-10T05:38:19.379 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-10T05:38:19.380 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-10T05:38:19.402 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean all 2026-03-10T05:38:19.492 INFO:teuthology.orchestra.run.vm05.stdout:56 files removed 2026-03-10T05:38:19.493 INFO:teuthology.orchestra.run.vm02.stdout:56 files removed 2026-03-10T05:38:19.516 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T05:38:19.518 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T05:38:19.538 INFO:teuthology.orchestra.run.vm01.stdout:56 files removed 2026-03-10T05:38:19.541 DEBUG:teuthology.orchestra.run.vm05:> sudo yum clean expire-cache 2026-03-10T05:38:19.545 DEBUG:teuthology.orchestra.run.vm02:> sudo yum clean expire-cache 2026-03-10T05:38:19.557 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T05:38:19.580 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean expire-cache 2026-03-10T05:38:19.693 INFO:teuthology.orchestra.run.vm05.stdout:Cache was expired 2026-03-10T05:38:19.693 INFO:teuthology.orchestra.run.vm05.stdout:0 files removed 2026-03-10T05:38:19.702 INFO:teuthology.orchestra.run.vm02.stdout:Cache was expired 2026-03-10T05:38:19.702 INFO:teuthology.orchestra.run.vm02.stdout:0 files removed 2026-03-10T05:38:19.715 DEBUG:teuthology.parallel:result is None 2026-03-10T05:38:19.724 DEBUG:teuthology.parallel:result is None 2026-03-10T05:38:19.735 INFO:teuthology.orchestra.run.vm01.stdout:Cache was expired 2026-03-10T05:38:19.735 INFO:teuthology.orchestra.run.vm01.stdout:0 files removed 2026-03-10T05:38:19.752 DEBUG:teuthology.parallel:result is None 2026-03-10T05:38:19.752 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm01.local 2026-03-10T05:38:19.752 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm02.local 2026-03-10T05:38:19.752 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm05.local 2026-03-10T05:38:19.752 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T05:38:19.752 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T05:38:19.752 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T05:38:19.777 DEBUG:teuthology.orchestra.run.vm05:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-10T05:38:19.779 DEBUG:teuthology.orchestra.run.vm02:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-10T05:38:19.779 DEBUG:teuthology.orchestra.run.vm01:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-10T05:38:19.844 DEBUG:teuthology.parallel:result is None 2026-03-10T05:38:19.846 DEBUG:teuthology.parallel:result is None 2026-03-10T05:38:19.848 DEBUG:teuthology.parallel:result is None 2026-03-10T05:38:19.848 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-10T05:38:19.850 INFO:tasks.cephadm:Teardown begin 2026-03-10T05:38:19.851 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:38:19.916 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:38:19.941 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:38:19.967 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-10T05:38:19.968 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba -- ceph mgr module disable cephadm 2026-03-10T05:38:20.139 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/mon.a/config 2026-03-10T05:38:20.160 INFO:teuthology.orchestra.run.vm01.stderr:Error: statfs /etc/ceph/ceph.client.admin.keyring: no such file or directory 2026-03-10T05:38:20.181 DEBUG:teuthology.orchestra.run:got remote process result: 125 2026-03-10T05:38:20.181 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-10T05:38:20.182 DEBUG:teuthology.orchestra.run.vm01:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T05:38:20.200 DEBUG:teuthology.orchestra.run.vm02:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T05:38:20.215 DEBUG:teuthology.orchestra.run.vm05:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T05:38:20.230 INFO:tasks.cephadm:Stopping all daemons... 2026-03-10T05:38:20.230 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-10T05:38:20.230 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.a 2026-03-10T05:38:20.268 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.a.service' 2026-03-10T05:38:20.337 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T05:38:20.337 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-10T05:38:20.337 INFO:tasks.cephadm.mon.c:Stopping mon.b... 2026-03-10T05:38:20.337 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.b 2026-03-10T05:38:20.365 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.b.service' 2026-03-10T05:38:20.435 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T05:38:20.435 INFO:tasks.cephadm.mon.c:Stopped mon.b 2026-03-10T05:38:20.436 INFO:tasks.cephadm.mon.c:Stopping mon.c... 2026-03-10T05:38:20.436 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.c 2026-03-10T05:38:20.465 DEBUG:teuthology.orchestra.run.vm05:> sudo pkill -f 'journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mon.c.service' 2026-03-10T05:38:20.539 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T05:38:20.539 INFO:tasks.cephadm.mon.c:Stopped mon.c 2026-03-10T05:38:20.539 INFO:tasks.cephadm.mgr.x:Stopping mgr.x... 2026-03-10T05:38:20.539 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mgr.x 2026-03-10T05:38:20.567 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@mgr.x.service' 2026-03-10T05:38:20.635 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T05:38:20.635 INFO:tasks.cephadm.mgr.x:Stopped mgr.x 2026-03-10T05:38:20.635 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-10T05:38:20.635 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.0 2026-03-10T05:38:20.703 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.0.service' 2026-03-10T05:38:20.773 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T05:38:20.773 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-10T05:38:20.773 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-10T05:38:20.773 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.1 2026-03-10T05:38:20.842 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.1.service' 2026-03-10T05:38:20.907 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T05:38:20.907 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-10T05:38:20.907 INFO:tasks.cephadm.osd.2:Stopping osd.2... 2026-03-10T05:38:20.907 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.2 2026-03-10T05:38:20.938 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.2.service' 2026-03-10T05:38:21.009 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T05:38:21.009 INFO:tasks.cephadm.osd.2:Stopped osd.2 2026-03-10T05:38:21.009 INFO:tasks.cephadm.osd.3:Stopping osd.3... 2026-03-10T05:38:21.009 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.3 2026-03-10T05:38:21.080 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.3.service' 2026-03-10T05:38:21.152 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T05:38:21.152 INFO:tasks.cephadm.osd.3:Stopped osd.3 2026-03-10T05:38:21.153 INFO:tasks.cephadm.osd.4:Stopping osd.4... 2026-03-10T05:38:21.153 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.4 2026-03-10T05:38:21.223 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.4.service' 2026-03-10T05:38:21.293 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T05:38:21.293 INFO:tasks.cephadm.osd.4:Stopped osd.4 2026-03-10T05:38:21.293 INFO:tasks.cephadm.osd.5:Stopping osd.5... 2026-03-10T05:38:21.293 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.5 2026-03-10T05:38:21.322 DEBUG:teuthology.orchestra.run.vm05:> sudo pkill -f 'journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.5.service' 2026-03-10T05:38:21.391 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T05:38:21.391 INFO:tasks.cephadm.osd.5:Stopped osd.5 2026-03-10T05:38:21.391 INFO:tasks.cephadm.osd.6:Stopping osd.6... 2026-03-10T05:38:21.391 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.6 2026-03-10T05:38:21.459 DEBUG:teuthology.orchestra.run.vm05:> sudo pkill -f 'journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.6.service' 2026-03-10T05:38:21.529 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T05:38:21.529 INFO:tasks.cephadm.osd.6:Stopped osd.6 2026-03-10T05:38:21.529 INFO:tasks.cephadm.osd.7:Stopping osd.7... 2026-03-10T05:38:21.529 DEBUG:teuthology.orchestra.run.vm05:> sudo systemctl stop ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.7 2026-03-10T05:38:21.599 DEBUG:teuthology.orchestra.run.vm05:> sudo pkill -f 'journalctl -f -n 0 -u ceph-dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba@osd.7.service' 2026-03-10T05:38:21.670 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T05:38:21.671 INFO:tasks.cephadm.osd.7:Stopped osd.7 2026-03-10T05:38:21.671 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba --force --keep-logs 2026-03-10T05:38:21.794 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:38:23.254 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba --force --keep-logs 2026-03-10T05:38:23.386 INFO:teuthology.orchestra.run.vm02.stdout:Deleting cluster with fsid: dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:38:24.577 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba --force --keep-logs 2026-03-10T05:38:24.713 INFO:teuthology.orchestra.run.vm05.stdout:Deleting cluster with fsid: dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:38:26.175 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:38:26.203 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:38:26.231 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T05:38:26.261 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-10T05:38:26.261 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/914/remote/vm01/crash 2026-03-10T05:38:26.261 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/crash -- . 2026-03-10T05:38:26.287 INFO:teuthology.orchestra.run.vm01.stderr:tar: /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/crash: Cannot open: No such file or directory 2026-03-10T05:38:26.287 INFO:teuthology.orchestra.run.vm01.stderr:tar: Error is not recoverable: exiting now 2026-03-10T05:38:26.288 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/914/remote/vm02/crash 2026-03-10T05:38:26.288 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/crash -- . 2026-03-10T05:38:26.312 INFO:teuthology.orchestra.run.vm02.stderr:tar: /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/crash: Cannot open: No such file or directory 2026-03-10T05:38:26.312 INFO:teuthology.orchestra.run.vm02.stderr:tar: Error is not recoverable: exiting now 2026-03-10T05:38:26.313 DEBUG:teuthology.misc:Transferring archived files from vm05:/var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/914/remote/vm05/crash 2026-03-10T05:38:26.313 DEBUG:teuthology.orchestra.run.vm05:> sudo tar c -f - -C /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/crash -- . 2026-03-10T05:38:26.341 INFO:teuthology.orchestra.run.vm05.stderr:tar: /var/lib/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/crash: Cannot open: No such file or directory 2026-03-10T05:38:26.342 INFO:teuthology.orchestra.run.vm05.stderr:tar: Error is not recoverable: exiting now 2026-03-10T05:38:26.342 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-10T05:38:26.343 DEBUG:teuthology.orchestra.run.vm01:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/ceph.log | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v MON_DOWN | head -n 1 2026-03-10T05:38:26.367 INFO:teuthology.orchestra.run.vm01.stderr:grep: /var/log/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/ceph.log: No such file or directory 2026-03-10T05:38:26.368 WARNING:tasks.cephadm:Found errors (ERR|WRN|SEC) in cluster log 2026-03-10T05:38:26.368 DEBUG:teuthology.orchestra.run.vm01:> sudo egrep '\[SEC\]' /var/log/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/ceph.log | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v MON_DOWN | head -n 1 2026-03-10T05:38:26.433 INFO:teuthology.orchestra.run.vm01.stderr:grep: /var/log/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/ceph.log: No such file or directory 2026-03-10T05:38:26.435 INFO:tasks.cephadm:Compressing logs... 2026-03-10T05:38:26.435 DEBUG:teuthology.orchestra.run.vm01:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T05:38:26.477 DEBUG:teuthology.orchestra.run.vm02:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T05:38:26.479 DEBUG:teuthology.orchestra.run.vm05:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T05:38:26.499 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T05:38:26.499 INFO:teuthology.orchestra.run.vm01.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T05:38:26.500 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/cephadm.log: 87.2% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T05:38:26.501 INFO:teuthology.orchestra.run.vm02.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T05:38:26.501 INFO:teuthology.orchestra.run.vm02.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T05:38:26.501 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-10T05:38:26.501 INFO:teuthology.orchestra.run.vm01.stderr:real 0m0.012s 2026-03-10T05:38:26.501 INFO:teuthology.orchestra.run.vm01.stderr:user 0m0.001s 2026-03-10T05:38:26.501 INFO:teuthology.orchestra.run.vm01.stderr:sys 0m0.017s 2026-03-10T05:38:26.502 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/cephadm.log: 89.3% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T05:38:26.503 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-10T05:38:26.503 INFO:teuthology.orchestra.run.vm02.stderr:real 0m0.012s 2026-03-10T05:38:26.503 INFO:teuthology.orchestra.run.vm02.stderr:user 0m0.005s 2026-03-10T05:38:26.503 INFO:teuthology.orchestra.run.vm02.stderr:sys 0m0.014s 2026-03-10T05:38:26.508 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T05:38:26.508 INFO:teuthology.orchestra.run.vm05.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T05:38:26.509 INFO:teuthology.orchestra.run.vm05.stderr:/var/log/ceph/cephadm.log: 90.1% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T05:38:26.510 INFO:teuthology.orchestra.run.vm05.stderr: 2026-03-10T05:38:26.510 INFO:teuthology.orchestra.run.vm05.stderr:real 0m0.015s 2026-03-10T05:38:26.510 INFO:teuthology.orchestra.run.vm05.stderr:user 0m0.007s 2026-03-10T05:38:26.510 INFO:teuthology.orchestra.run.vm05.stderr:sys 0m0.014s 2026-03-10T05:38:26.511 INFO:tasks.cephadm:Archiving logs... 2026-03-10T05:38:26.511 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/914/remote/vm01/log 2026-03-10T05:38:26.511 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T05:38:26.565 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/914/remote/vm02/log 2026-03-10T05:38:26.565 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T05:38:26.589 DEBUG:teuthology.misc:Transferring archived files from vm05:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/914/remote/vm05/log 2026-03-10T05:38:26.589 DEBUG:teuthology.orchestra.run.vm05:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T05:38:26.613 INFO:tasks.cephadm:Removing cluster... 2026-03-10T05:38:26.613 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba --force 2026-03-10T05:38:26.740 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:38:26.834 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba --force 2026-03-10T05:38:26.960 INFO:teuthology.orchestra.run.vm02.stdout:Deleting cluster with fsid: dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:38:27.062 DEBUG:teuthology.orchestra.run.vm05:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba --force 2026-03-10T05:38:27.189 INFO:teuthology.orchestra.run.vm05.stdout:Deleting cluster with fsid: dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba 2026-03-10T05:38:27.290 INFO:tasks.cephadm:Removing cephadm ... 2026-03-10T05:38:27.290 DEBUG:teuthology.orchestra.run.vm01:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T05:38:27.308 DEBUG:teuthology.orchestra.run.vm02:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T05:38:27.323 DEBUG:teuthology.orchestra.run.vm05:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T05:38:27.342 INFO:tasks.cephadm:Teardown complete 2026-03-10T05:38:27.342 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-10T05:38:27.345 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-10T05:38:27.345 DEBUG:teuthology.orchestra.run.vm01:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T05:38:27.350 DEBUG:teuthology.orchestra.run.vm02:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T05:38:27.362 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-10T05:38:27.364 DEBUG:teuthology.orchestra.run.vm05:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T05:38:27.377 INFO:teuthology.orchestra.run.vm02.stderr:bash: line 1: ntpq: command not found 2026-03-10T05:38:27.400 INFO:teuthology.orchestra.run.vm05.stderr:bash: line 1: ntpq: command not found 2026-03-10T05:38:27.558 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T05:38:27.558 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-10T05:38:27.558 INFO:teuthology.orchestra.run.vm01.stdout:^- www.festibal.de 2 6 377 57 +148us[ +146us] +/- 46ms 2026-03-10T05:38:27.558 INFO:teuthology.orchestra.run.vm01.stdout:^+ home.of.the.smiling-prox> 1 6 377 25 -2352us[-2352us] +/- 20ms 2026-03-10T05:38:27.558 INFO:teuthology.orchestra.run.vm01.stdout:^* mail.anyvm.tech 2 6 377 56 -59us[ -61us] +/- 16ms 2026-03-10T05:38:27.559 INFO:teuthology.orchestra.run.vm01.stdout:^+ 139-162-156-95.ip.linode> 2 6 377 55 +4389us[+4389us] +/- 34ms 2026-03-10T05:38:27.559 INFO:teuthology.orchestra.run.vm05.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T05:38:27.559 INFO:teuthology.orchestra.run.vm05.stdout:=============================================================================== 2026-03-10T05:38:27.559 INFO:teuthology.orchestra.run.vm05.stdout:^+ 139-162-156-95.ip.linode> 2 6 377 60 +2941us[+2940us] +/- 34ms 2026-03-10T05:38:27.559 INFO:teuthology.orchestra.run.vm05.stdout:^- www.festibal.de 2 6 377 59 -1284us[-1285us] +/- 46ms 2026-03-10T05:38:27.559 INFO:teuthology.orchestra.run.vm05.stdout:^* mail.anyvm.tech 2 6 377 57 -1417us[-1418us] +/- 16ms 2026-03-10T05:38:27.559 INFO:teuthology.orchestra.run.vm05.stdout:^- router02.i-tk.de 2 6 377 30 -153us[ -153us] +/- 25ms 2026-03-10T05:38:27.560 INFO:teuthology.orchestra.run.vm02.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T05:38:27.560 INFO:teuthology.orchestra.run.vm02.stdout:=============================================================================== 2026-03-10T05:38:27.560 INFO:teuthology.orchestra.run.vm02.stdout:^+ 139-162-156-95.ip.linode> 2 6 377 59 +3903us[+3896us] +/- 34ms 2026-03-10T05:38:27.560 INFO:teuthology.orchestra.run.vm02.stdout:^- time.cloudflare.com 3 6 377 39 -2688us[-2688us] +/- 15ms 2026-03-10T05:38:27.560 INFO:teuthology.orchestra.run.vm02.stdout:^- www.festibal.de 2 6 377 58 -1205us[-1211us] +/- 46ms 2026-03-10T05:38:27.560 INFO:teuthology.orchestra.run.vm02.stdout:^* mail.anyvm.tech 2 6 377 57 -1154us[-1160us] +/- 16ms 2026-03-10T05:38:27.560 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-10T05:38:27.562 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-10T05:38:27.563 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-10T05:38:27.565 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-10T05:38:27.567 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-10T05:38:27.569 INFO:teuthology.task.internal:Duration was 763.684236 seconds 2026-03-10T05:38:27.569 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-10T05:38:27.571 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-10T05:38:27.571 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T05:38:27.601 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T05:38:27.602 DEBUG:teuthology.orchestra.run.vm05:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T05:38:27.635 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T05:38:27.636 INFO:teuthology.orchestra.run.vm02.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T05:38:27.644 INFO:teuthology.orchestra.run.vm05.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T05:38:28.112 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-10T05:38:28.112 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm01.local 2026-03-10T05:38:28.112 DEBUG:teuthology.orchestra.run.vm01:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T05:38:28.174 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm02.local 2026-03-10T05:38:28.174 DEBUG:teuthology.orchestra.run.vm02:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T05:38:28.198 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm05.local 2026-03-10T05:38:28.198 DEBUG:teuthology.orchestra.run.vm05:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T05:38:28.223 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-10T05:38:28.223 DEBUG:teuthology.orchestra.run.vm01:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T05:38:28.224 DEBUG:teuthology.orchestra.run.vm02:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T05:38:28.240 DEBUG:teuthology.orchestra.run.vm05:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T05:38:28.666 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-10T05:38:28.666 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T05:38:28.668 DEBUG:teuthology.orchestra.run.vm02:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T05:38:28.669 DEBUG:teuthology.orchestra.run.vm05:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T05:38:28.694 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T05:38:28.694 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T05:38:28.695 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T05:38:28.695 INFO:teuthology.orchestra.run.vm02.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: gzip 0.0% -5 -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T05:38:28.695 INFO:teuthology.orchestra.run.vm02.stderr: --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T05:38:28.695 INFO:teuthology.orchestra.run.vm02.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz/home/ubuntu/cephtest/archive/syslog/journalctl.log: 2026-03-10T05:38:28.695 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T05:38:28.695 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 0.0% 2026-03-10T05:38:28.695 INFO:teuthology.orchestra.run.vm01.stderr: -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T05:38:28.696 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T05:38:28.697 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T05:38:28.697 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T05:38:28.697 INFO:teuthology.orchestra.run.vm05.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T05:38:28.697 INFO:teuthology.orchestra.run.vm05.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T05:38:28.698 INFO:teuthology.orchestra.run.vm05.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: /home/ubuntu/cephtest/archive/syslog/journalctl.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T05:38:28.819 INFO:teuthology.orchestra.run.vm02.stderr: 98.3% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T05:38:28.836 INFO:teuthology.orchestra.run.vm05.stderr: 98.3% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T05:38:28.839 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.2% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T05:38:28.841 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-10T05:38:28.844 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-10T05:38:28.844 DEBUG:teuthology.orchestra.run.vm01:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T05:38:28.910 DEBUG:teuthology.orchestra.run.vm02:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T05:38:28.938 DEBUG:teuthology.orchestra.run.vm05:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T05:38:28.963 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-10T05:38:28.965 DEBUG:teuthology.orchestra.run.vm01:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T05:38:28.967 DEBUG:teuthology.orchestra.run.vm02:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T05:38:28.980 DEBUG:teuthology.orchestra.run.vm05:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T05:38:28.990 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = core 2026-03-10T05:38:29.005 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = core 2026-03-10T05:38:29.032 INFO:teuthology.orchestra.run.vm05.stdout:kernel.core_pattern = core 2026-03-10T05:38:29.046 DEBUG:teuthology.orchestra.run.vm01:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T05:38:29.062 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:38:29.062 DEBUG:teuthology.orchestra.run.vm02:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T05:38:29.080 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:38:29.080 DEBUG:teuthology.orchestra.run.vm05:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T05:38:29.100 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T05:38:29.100 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-10T05:38:29.103 INFO:teuthology.task.internal:Transferring archived files... 2026-03-10T05:38:29.104 DEBUG:teuthology.misc:Transferring archived files from vm01:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/914/remote/vm01 2026-03-10T05:38:29.104 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T05:38:29.136 DEBUG:teuthology.misc:Transferring archived files from vm02:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/914/remote/vm02 2026-03-10T05:38:29.136 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T05:38:29.165 DEBUG:teuthology.misc:Transferring archived files from vm05:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/914/remote/vm05 2026-03-10T05:38:29.166 DEBUG:teuthology.orchestra.run.vm05:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T05:38:29.192 INFO:teuthology.task.internal:Removing archive directory... 2026-03-10T05:38:29.192 DEBUG:teuthology.orchestra.run.vm01:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T05:38:29.194 DEBUG:teuthology.orchestra.run.vm02:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T05:38:29.207 DEBUG:teuthology.orchestra.run.vm05:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T05:38:29.246 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-10T05:38:29.249 INFO:teuthology.task.internal:Not uploading archives. 2026-03-10T05:38:29.249 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-10T05:38:29.251 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-10T05:38:29.251 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T05:38:29.253 DEBUG:teuthology.orchestra.run.vm02:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T05:38:29.264 DEBUG:teuthology.orchestra.run.vm05:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T05:38:29.268 INFO:teuthology.orchestra.run.vm01.stdout: 8532140 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 10 05:38 /home/ubuntu/cephtest 2026-03-10T05:38:29.279 INFO:teuthology.orchestra.run.vm02.stdout: 8532145 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 10 05:38 /home/ubuntu/cephtest 2026-03-10T05:38:29.301 INFO:teuthology.orchestra.run.vm05.stdout: 8532144 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 10 05:38 /home/ubuntu/cephtest 2026-03-10T05:38:29.302 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-10T05:38:29.308 INFO:teuthology.run:Summary data: description: orch/cephadm/rbd_iscsi/{base/install cluster/{fixed-3 openstack} conf/{disable-pool-app} supported-container-hosts$/{centos_9.stream_runc} workloads/cephadm_iscsi} duration: 763.6842355728149 failure_reason: '"grep: /var/log/ceph/dbc0fe1e-1c41-11f1-a82d-e7b0f6c6f3ba/ceph.log: No such file or directory" in cluster log' flavor: default owner: kyr success: false 2026-03-10T05:38:29.308 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T05:38:29.333 INFO:teuthology.run:FAIL